2026-03-10T10:17:34.633 INFO:root:teuthology version: 1.2.4.dev6+g1c580df7a 2026-03-10T10:17:34.639 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-10T10:17:34.666 INFO:teuthology.run:Config: archive_path: /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/998 branch: squid description: orch/cephadm/mgr-nfs-upgrade/{0-centos_9.stream 1-bootstrap/17.2.0 1-start 2-nfs 3-upgrade-with-workload 4-final} email: null first_in_suite: false flavor: default job_id: '998' last_in_suite: false machine_type: vps name: kyr-2026-03-10_01:00:38-orch-squid-none-default-vps no_nested_subset: false openstack: - volumes: count: 4 size: 10 os_type: centos os_version: 9.stream overrides: admin_socket: branch: squid ansible.cephlab: branch: main skip_tags: nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs vars: timezone: UTC ceph: conf: mgr: debug mgr: 20 debug ms: 1 mon: debug mon: 20 debug ms: 1 debug paxos: 20 osd: debug ms: 1 debug osd: 20 osd mclock iops capacity threshold hdd: 49000 osd shutdown pgref assert: true flavor: default log-ignorelist: - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - CEPHADM_REFRESH_FAILED log-only-match: - CEPHADM_ sha1: e911bdebe5c8faa3800735d1568fcdca65db60df ceph-deploy: conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: {} install: ceph: flavor: default sha1: e911bdebe5c8faa3800735d1568fcdca65db60df extra_system_packages: deb: - python3-xmltodict - python3-jmespath rpm: - bzip2 - perl-Test-Harness - python3-xmltodict - python3-jmespath selinux: allowlist: - scontext=system_u:system_r:logrotate_t:s0 - scontext=system_u:system_r:getty_t:s0 workunit: branch: tt-squid sha1: 75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b owner: kyr priority: 1000 repo: https://github.com/ceph/ceph.git roles: - - host.a - osd.0 - osd.1 - osd.2 - osd.3 - client.0 - - host.b - osd.4 - osd.5 - osd.6 - osd.7 seed: 8043 sha1: e911bdebe5c8faa3800735d1568fcdca65db60df sleep_before_teardown: 0 subset: 1/64 suite: orch suite_branch: tt-squid suite_path: /home/teuthos/src/github.com_kshtsk_ceph_75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b/qa suite_relpath: qa suite_repo: https://github.com/kshtsk/ceph.git suite_sha1: 75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b targets: vm01.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBIB3b62HvfiHCT3NIL2CQH8mjNM0kZvGR0BnQYF6SL7eodqPHGYFs74mB/ixlUQPurYhTqbnDOJFhyS1a+mSsHs= vm08.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBFKmKa0uHJuHAaub56X2bIQAXIIcDdeFpOYhsfcbZY+R2a5f56HDw6DpzC9G952xXXwtd1b4KKd8PY6nruUo/2Q= tasks: - cephadm: cephadm_branch: v17.2.0 cephadm_git_url: https://github.com/ceph/ceph image: quay.io/ceph/ceph:v17.2.0 roleless: true - cephadm.shell: host.a: - ceph orch status - ceph orch ps - ceph orch ls - ceph orch host ls - ceph orch device ls - vip.exec: all-hosts: - systemctl stop nfs-server - cephadm.shell: host.a: - ceph fs volume create foofs - cephadm.wait_for_service: service: mds.foofs - cephadm.shell: host.a: - ceph nfs cluster create foo --placement=2 || ceph nfs cluster create cephfs foo --placement=2 - ceph nfs export create cephfs --fsname foofs --clusterid foo --binding /fake || ceph nfs export create cephfs --fsname foofs --cluster-id foo --pseudo-path /fake - while ! ceph orch ls | grep nfs | grep 2/2 ; do sleep 1 ; done - vip.exec: host.a: - mkdir /mnt/foo - while ! mount -t nfs $(hostname):/fake /mnt/foo -o sync ; do sleep 5 ; done - echo test > /mnt/foo/testfile - sync - parallel: - upgrade-tasks - workload-tasks - vip.exec: host.a: - umount /mnt/foo - cephadm.shell: host.a: - ceph nfs cluster ls | grep foo - ceph nfs export ls foo --detailed - rados -p .nfs --all ls - - cephadm.shell: host.a: - 'set -ex [[ `ceph config get mgr mgr/cephadm/migration_current` -gt 2 ]] ' teuthology: fragments_dropped: [] meta: {} postmerge: [] teuthology_branch: clyso-debian-13 teuthology_repo: https://github.com/clyso/teuthology teuthology_sha1: 1c580df7a9c7c2aadc272da296344fd99f27c444 timestamp: 2026-03-10_01:00:38 tube: vps upgrade-tasks: sequential: - cephadm.shell: env: - sha1 host.a: - ceph config set mon mon_warn_on_insecure_global_id_reclaim false --force - ceph config set mon mon_warn_on_insecure_global_id_reclaim_allowed false --force - ceph config set global log_to_journald false --force - ceph mgr module enable nfs --force - ceph orch upgrade start --image quay.ceph.io/ceph-ci/ceph:$sha1 - cephadm.shell: env: - sha1 host.a: - while ceph orch upgrade status | jq '.in_progress' | grep true && ! ceph orch upgrade status | jq '.message' | grep Error ; do ceph orch ps ; ceph versions ; ceph orch upgrade status ; ceph health detail ; sleep 30 ; done - ceph orch ps - ceph versions - echo "wait for servicemap items w/ changing names to refresh" - sleep 60 - ceph orch ps - ceph orch upgrade status - ceph health detail - ceph versions - ceph versions | jq -e '.overall | length == 1' - ceph versions | jq -e '.overall | keys' | grep $sha1 - cephadm.wait_for_service: service: nfs.foo user: kyr verbose: false worker_log: /home/teuthos/.teuthology/dispatcher/dispatcher.vps.611473 workload-tasks: sequential: - exec: host.a: - cd /mnt/foo && dbench 5 -t 600 || true - umount /mnt/foo - while ! mount -t nfs $(hostname):/fake /mnt/foo ; do sleep 5 ; done - cd /mnt/foo && dbench 5 -t 5 2026-03-10T10:17:34.666 INFO:teuthology.run:suite_path is set to /home/teuthos/src/github.com_kshtsk_ceph_75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b/qa; will attempt to use it 2026-03-10T10:17:34.666 INFO:teuthology.run:Found tasks at /home/teuthos/src/github.com_kshtsk_ceph_75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b/qa/tasks 2026-03-10T10:17:34.667 INFO:teuthology.run_tasks:Running task internal.check_packages... 2026-03-10T10:17:34.667 INFO:teuthology.task.internal:Checking packages... 2026-03-10T10:17:34.667 INFO:teuthology.task.internal:Checking packages for os_type 'centos', flavor 'default' and ceph hash 'e911bdebe5c8faa3800735d1568fcdca65db60df' 2026-03-10T10:17:34.667 WARNING:teuthology.packaging:More than one of ref, tag, branch, or sha1 supplied; using branch 2026-03-10T10:17:34.667 INFO:teuthology.packaging:ref: None 2026-03-10T10:17:34.667 INFO:teuthology.packaging:tag: None 2026-03-10T10:17:34.667 INFO:teuthology.packaging:branch: squid 2026-03-10T10:17:34.667 INFO:teuthology.packaging:sha1: e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T10:17:34.667 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&ref=squid 2026-03-10T10:17:35.442 INFO:teuthology.task.internal:Found packages for ceph version 19.2.3-678.ge911bdeb 2026-03-10T10:17:35.443 INFO:teuthology.run_tasks:Running task internal.buildpackages_prep... 2026-03-10T10:17:35.444 INFO:teuthology.task.internal:no buildpackages task found 2026-03-10T10:17:35.444 INFO:teuthology.run_tasks:Running task internal.save_config... 2026-03-10T10:17:35.445 INFO:teuthology.task.internal:Saving configuration 2026-03-10T10:17:35.450 INFO:teuthology.run_tasks:Running task internal.check_lock... 2026-03-10T10:17:35.451 INFO:teuthology.task.internal.check_lock:Checking locks... 2026-03-10T10:17:35.457 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm01.local', 'description': '/archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/998', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-10 10:16:32.967948', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:01', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBIB3b62HvfiHCT3NIL2CQH8mjNM0kZvGR0BnQYF6SL7eodqPHGYFs74mB/ixlUQPurYhTqbnDOJFhyS1a+mSsHs='} 2026-03-10T10:17:35.462 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm08.local', 'description': '/archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/998', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-10 10:16:32.968378', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:08', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBFKmKa0uHJuHAaub56X2bIQAXIIcDdeFpOYhsfcbZY+R2a5f56HDw6DpzC9G952xXXwtd1b4KKd8PY6nruUo/2Q='} 2026-03-10T10:17:35.462 INFO:teuthology.run_tasks:Running task internal.add_remotes... 2026-03-10T10:17:35.463 INFO:teuthology.task.internal:roles: ubuntu@vm01.local - ['host.a', 'osd.0', 'osd.1', 'osd.2', 'osd.3', 'client.0'] 2026-03-10T10:17:35.463 INFO:teuthology.task.internal:roles: ubuntu@vm08.local - ['host.b', 'osd.4', 'osd.5', 'osd.6', 'osd.7'] 2026-03-10T10:17:35.463 INFO:teuthology.run_tasks:Running task console_log... 2026-03-10T10:17:35.469 DEBUG:teuthology.task.console_log:vm01 does not support IPMI; excluding 2026-03-10T10:17:35.475 DEBUG:teuthology.task.console_log:vm08 does not support IPMI; excluding 2026-03-10T10:17:35.475 DEBUG:teuthology.exit:Installing handler: Handler(exiter=, func=.kill_console_loggers at 0x7fda6617e170>, signals=[15]) 2026-03-10T10:17:35.475 INFO:teuthology.run_tasks:Running task internal.connect... 2026-03-10T10:17:35.476 INFO:teuthology.task.internal:Opening connections... 2026-03-10T10:17:35.476 DEBUG:teuthology.task.internal:connecting to ubuntu@vm01.local 2026-03-10T10:17:35.477 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm01.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T10:17:35.536 DEBUG:teuthology.task.internal:connecting to ubuntu@vm08.local 2026-03-10T10:17:35.536 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm08.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T10:17:35.595 INFO:teuthology.run_tasks:Running task internal.push_inventory... 2026-03-10T10:17:35.596 DEBUG:teuthology.orchestra.run.vm01:> uname -m 2026-03-10T10:17:35.654 INFO:teuthology.orchestra.run.vm01.stdout:x86_64 2026-03-10T10:17:35.654 DEBUG:teuthology.orchestra.run.vm01:> cat /etc/os-release 2026-03-10T10:17:35.710 INFO:teuthology.orchestra.run.vm01.stdout:NAME="CentOS Stream" 2026-03-10T10:17:35.710 INFO:teuthology.orchestra.run.vm01.stdout:VERSION="9" 2026-03-10T10:17:35.710 INFO:teuthology.orchestra.run.vm01.stdout:ID="centos" 2026-03-10T10:17:35.710 INFO:teuthology.orchestra.run.vm01.stdout:ID_LIKE="rhel fedora" 2026-03-10T10:17:35.710 INFO:teuthology.orchestra.run.vm01.stdout:VERSION_ID="9" 2026-03-10T10:17:35.710 INFO:teuthology.orchestra.run.vm01.stdout:PLATFORM_ID="platform:el9" 2026-03-10T10:17:35.710 INFO:teuthology.orchestra.run.vm01.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-10T10:17:35.711 INFO:teuthology.orchestra.run.vm01.stdout:ANSI_COLOR="0;31" 2026-03-10T10:17:35.711 INFO:teuthology.orchestra.run.vm01.stdout:LOGO="fedora-logo-icon" 2026-03-10T10:17:35.711 INFO:teuthology.orchestra.run.vm01.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-10T10:17:35.711 INFO:teuthology.orchestra.run.vm01.stdout:HOME_URL="https://centos.org/" 2026-03-10T10:17:35.711 INFO:teuthology.orchestra.run.vm01.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-10T10:17:35.711 INFO:teuthology.orchestra.run.vm01.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-10T10:17:35.711 INFO:teuthology.orchestra.run.vm01.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-10T10:17:35.711 INFO:teuthology.lock.ops:Updating vm01.local on lock server 2026-03-10T10:17:35.715 DEBUG:teuthology.orchestra.run.vm08:> uname -m 2026-03-10T10:17:35.730 INFO:teuthology.orchestra.run.vm08.stdout:x86_64 2026-03-10T10:17:35.730 DEBUG:teuthology.orchestra.run.vm08:> cat /etc/os-release 2026-03-10T10:17:35.786 INFO:teuthology.orchestra.run.vm08.stdout:NAME="CentOS Stream" 2026-03-10T10:17:35.786 INFO:teuthology.orchestra.run.vm08.stdout:VERSION="9" 2026-03-10T10:17:35.786 INFO:teuthology.orchestra.run.vm08.stdout:ID="centos" 2026-03-10T10:17:35.786 INFO:teuthology.orchestra.run.vm08.stdout:ID_LIKE="rhel fedora" 2026-03-10T10:17:35.786 INFO:teuthology.orchestra.run.vm08.stdout:VERSION_ID="9" 2026-03-10T10:17:35.786 INFO:teuthology.orchestra.run.vm08.stdout:PLATFORM_ID="platform:el9" 2026-03-10T10:17:35.786 INFO:teuthology.orchestra.run.vm08.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-10T10:17:35.786 INFO:teuthology.orchestra.run.vm08.stdout:ANSI_COLOR="0;31" 2026-03-10T10:17:35.786 INFO:teuthology.orchestra.run.vm08.stdout:LOGO="fedora-logo-icon" 2026-03-10T10:17:35.786 INFO:teuthology.orchestra.run.vm08.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-10T10:17:35.786 INFO:teuthology.orchestra.run.vm08.stdout:HOME_URL="https://centos.org/" 2026-03-10T10:17:35.786 INFO:teuthology.orchestra.run.vm08.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-10T10:17:35.786 INFO:teuthology.orchestra.run.vm08.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-10T10:17:35.786 INFO:teuthology.orchestra.run.vm08.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-10T10:17:35.786 INFO:teuthology.lock.ops:Updating vm08.local on lock server 2026-03-10T10:17:35.791 INFO:teuthology.run_tasks:Running task internal.serialize_remote_roles... 2026-03-10T10:17:35.793 INFO:teuthology.run_tasks:Running task internal.check_conflict... 2026-03-10T10:17:35.794 INFO:teuthology.task.internal:Checking for old test directory... 2026-03-10T10:17:35.794 DEBUG:teuthology.orchestra.run.vm01:> test '!' -e /home/ubuntu/cephtest 2026-03-10T10:17:35.796 DEBUG:teuthology.orchestra.run.vm08:> test '!' -e /home/ubuntu/cephtest 2026-03-10T10:17:35.840 INFO:teuthology.run_tasks:Running task internal.check_ceph_data... 2026-03-10T10:17:35.842 INFO:teuthology.task.internal:Checking for non-empty /var/lib/ceph... 2026-03-10T10:17:35.842 DEBUG:teuthology.orchestra.run.vm01:> test -z $(ls -A /var/lib/ceph) 2026-03-10T10:17:35.852 DEBUG:teuthology.orchestra.run.vm08:> test -z $(ls -A /var/lib/ceph) 2026-03-10T10:17:35.866 INFO:teuthology.orchestra.run.vm01.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-10T10:17:35.896 INFO:teuthology.orchestra.run.vm08.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-10T10:17:35.897 INFO:teuthology.run_tasks:Running task internal.vm_setup... 2026-03-10T10:17:35.905 DEBUG:teuthology.orchestra.run.vm01:> test -e /ceph-qa-ready 2026-03-10T10:17:35.920 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T10:17:36.107 DEBUG:teuthology.orchestra.run.vm08:> test -e /ceph-qa-ready 2026-03-10T10:17:36.120 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T10:17:36.398 INFO:teuthology.run_tasks:Running task internal.base... 2026-03-10T10:17:36.408 INFO:teuthology.task.internal:Creating test directory... 2026-03-10T10:17:36.409 DEBUG:teuthology.orchestra.run.vm01:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-10T10:17:36.411 DEBUG:teuthology.orchestra.run.vm08:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-10T10:17:36.427 INFO:teuthology.run_tasks:Running task internal.archive_upload... 2026-03-10T10:17:36.446 INFO:teuthology.run_tasks:Running task internal.archive... 2026-03-10T10:17:36.460 INFO:teuthology.task.internal:Creating archive directory... 2026-03-10T10:17:36.460 DEBUG:teuthology.orchestra.run.vm01:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-10T10:17:36.470 DEBUG:teuthology.orchestra.run.vm08:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-10T10:17:36.494 INFO:teuthology.run_tasks:Running task internal.coredump... 2026-03-10T10:17:36.495 INFO:teuthology.task.internal:Enabling coredump saving... 2026-03-10T10:17:36.496 DEBUG:teuthology.orchestra.run.vm01:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-10T10:17:36.540 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T10:17:36.540 DEBUG:teuthology.orchestra.run.vm08:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-10T10:17:36.555 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T10:17:36.555 DEBUG:teuthology.orchestra.run.vm01:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-10T10:17:36.581 DEBUG:teuthology.orchestra.run.vm08:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-10T10:17:36.604 INFO:teuthology.orchestra.run.vm01.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T10:17:36.614 INFO:teuthology.orchestra.run.vm01.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T10:17:36.621 INFO:teuthology.orchestra.run.vm08.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T10:17:36.629 INFO:teuthology.orchestra.run.vm08.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T10:17:36.630 INFO:teuthology.run_tasks:Running task internal.sudo... 2026-03-10T10:17:36.632 INFO:teuthology.task.internal:Configuring sudo... 2026-03-10T10:17:36.632 DEBUG:teuthology.orchestra.run.vm01:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-10T10:17:36.658 DEBUG:teuthology.orchestra.run.vm08:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-10T10:17:36.696 INFO:teuthology.run_tasks:Running task internal.syslog... 2026-03-10T10:17:36.714 INFO:teuthology.task.internal.syslog:Starting syslog monitoring... 2026-03-10T10:17:36.714 DEBUG:teuthology.orchestra.run.vm01:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-10T10:17:36.730 DEBUG:teuthology.orchestra.run.vm08:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-10T10:17:36.753 DEBUG:teuthology.orchestra.run.vm01:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T10:17:36.805 DEBUG:teuthology.orchestra.run.vm01:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T10:17:36.860 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T10:17:36.860 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-10T10:17:36.918 DEBUG:teuthology.orchestra.run.vm08:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T10:17:36.942 DEBUG:teuthology.orchestra.run.vm08:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T10:17:37.001 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-10T10:17:37.001 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-10T10:17:37.062 DEBUG:teuthology.orchestra.run.vm01:> sudo service rsyslog restart 2026-03-10T10:17:37.064 DEBUG:teuthology.orchestra.run.vm08:> sudo service rsyslog restart 2026-03-10T10:17:37.095 INFO:teuthology.orchestra.run.vm01.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-10T10:17:37.132 INFO:teuthology.orchestra.run.vm08.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-10T10:17:37.396 INFO:teuthology.run_tasks:Running task internal.timer... 2026-03-10T10:17:37.398 INFO:teuthology.task.internal:Starting timer... 2026-03-10T10:17:37.398 INFO:teuthology.run_tasks:Running task pcp... 2026-03-10T10:17:37.401 INFO:teuthology.run_tasks:Running task selinux... 2026-03-10T10:17:37.404 DEBUG:teuthology.task:Applying overrides for task selinux: {'allowlist': ['scontext=system_u:system_r:logrotate_t:s0', 'scontext=system_u:system_r:getty_t:s0']} 2026-03-10T10:17:37.404 INFO:teuthology.task.selinux:Excluding vm01: VMs are not yet supported 2026-03-10T10:17:37.404 INFO:teuthology.task.selinux:Excluding vm08: VMs are not yet supported 2026-03-10T10:17:37.404 DEBUG:teuthology.task.selinux:Getting current SELinux state 2026-03-10T10:17:37.404 DEBUG:teuthology.task.selinux:Existing SELinux modes: {} 2026-03-10T10:17:37.404 INFO:teuthology.task.selinux:Putting SELinux into permissive mode 2026-03-10T10:17:37.404 INFO:teuthology.run_tasks:Running task ansible.cephlab... 2026-03-10T10:17:37.406 DEBUG:teuthology.task:Applying overrides for task ansible.cephlab: {'branch': 'main', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'timezone': 'UTC'}} 2026-03-10T10:17:37.406 DEBUG:teuthology.repo_utils:Setting repo remote to https://github.com/ceph/ceph-cm-ansible.git 2026-03-10T10:17:37.407 INFO:teuthology.repo_utils:Fetching github.com_ceph_ceph-cm-ansible_main from origin 2026-03-10T10:17:37.887 DEBUG:teuthology.repo_utils:Resetting repo at /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main to origin/main 2026-03-10T10:17:37.892 INFO:teuthology.task.ansible:Playbook: [{'import_playbook': 'ansible_managed.yml'}, {'import_playbook': 'teuthology.yml'}, {'hosts': 'testnodes', 'tasks': [{'set_fact': {'ran_from_cephlab_playbook': True}}]}, {'import_playbook': 'testnodes.yml'}, {'import_playbook': 'container-host.yml'}, {'import_playbook': 'cobbler.yml'}, {'import_playbook': 'paddles.yml'}, {'import_playbook': 'pulpito.yml'}, {'hosts': 'testnodes', 'become': True, 'tasks': [{'name': 'Touch /ceph-qa-ready', 'file': {'path': '/ceph-qa-ready', 'state': 'touch'}, 'when': 'ran_from_cephlab_playbook|bool'}]}] 2026-03-10T10:17:37.892 DEBUG:teuthology.task.ansible:Running ansible-playbook -v --extra-vars '{"ansible_ssh_user": "ubuntu", "timezone": "UTC"}' -i /tmp/teuth_ansible_inventoryyn1gtmw2 --limit vm01.local,vm08.local /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main/cephlab.yml --skip-tags nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs 2026-03-10T10:19:40.774 DEBUG:teuthology.task.ansible:Reconnecting to [Remote(name='ubuntu@vm01.local'), Remote(name='ubuntu@vm08.local')] 2026-03-10T10:19:40.775 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm01.local' 2026-03-10T10:19:40.775 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm01.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T10:19:40.854 DEBUG:teuthology.orchestra.run.vm01:> true 2026-03-10T10:19:40.932 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm01.local' 2026-03-10T10:19:40.932 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm08.local' 2026-03-10T10:19:40.932 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm08.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T10:19:41.002 DEBUG:teuthology.orchestra.run.vm08:> true 2026-03-10T10:19:41.096 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm08.local' 2026-03-10T10:19:41.096 INFO:teuthology.run_tasks:Running task clock... 2026-03-10T10:19:41.119 INFO:teuthology.task.clock:Syncing clocks and checking initial clock skew... 2026-03-10T10:19:41.119 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-10T10:19:41.119 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T10:19:41.121 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-10T10:19:41.121 DEBUG:teuthology.orchestra.run.vm08:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T10:19:41.166 INFO:teuthology.orchestra.run.vm01.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-10T10:19:41.174 INFO:teuthology.orchestra.run.vm08.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-10T10:19:41.187 INFO:teuthology.orchestra.run.vm01.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-10T10:19:41.193 INFO:teuthology.orchestra.run.vm08.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-10T10:19:41.226 INFO:teuthology.orchestra.run.vm01.stderr:sudo: ntpd: command not found 2026-03-10T10:19:41.226 INFO:teuthology.orchestra.run.vm08.stderr:sudo: ntpd: command not found 2026-03-10T10:19:41.239 INFO:teuthology.orchestra.run.vm08.stdout:506 Cannot talk to daemon 2026-03-10T10:19:41.243 INFO:teuthology.orchestra.run.vm01.stdout:506 Cannot talk to daemon 2026-03-10T10:19:41.257 INFO:teuthology.orchestra.run.vm08.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-10T10:19:41.268 INFO:teuthology.orchestra.run.vm01.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-10T10:19:41.276 INFO:teuthology.orchestra.run.vm08.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-10T10:19:41.288 INFO:teuthology.orchestra.run.vm01.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-10T10:19:41.342 INFO:teuthology.orchestra.run.vm08.stderr:bash: line 1: ntpq: command not found 2026-03-10T10:19:41.345 INFO:teuthology.orchestra.run.vm08.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-10T10:19:41.345 INFO:teuthology.orchestra.run.vm08.stdout:=============================================================================== 2026-03-10T10:19:41.353 INFO:teuthology.orchestra.run.vm01.stderr:bash: line 1: ntpq: command not found 2026-03-10T10:19:41.355 INFO:teuthology.orchestra.run.vm01.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-10T10:19:41.356 INFO:teuthology.orchestra.run.vm01.stdout:=============================================================================== 2026-03-10T10:19:41.356 INFO:teuthology.run_tasks:Running task cephadm... 2026-03-10T10:19:41.411 INFO:tasks.cephadm:Config: {'cephadm_branch': 'v17.2.0', 'cephadm_git_url': 'https://github.com/ceph/ceph', 'image': 'quay.io/ceph/ceph:v17.2.0', 'roleless': True, 'conf': {'mgr': {'debug mgr': 20, 'debug ms': 1}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000, 'osd shutdown pgref assert': True}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', 'CEPHADM_REFRESH_FAILED'], 'log-only-match': ['CEPHADM_'], 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'} 2026-03-10T10:19:41.411 INFO:tasks.cephadm:Cluster image is quay.io/ceph/ceph:v17.2.0 2026-03-10T10:19:41.411 INFO:tasks.cephadm:Cluster fsid is a6b3e19a-1c6a-11f1-aaba-ff800f171c2f 2026-03-10T10:19:41.411 INFO:tasks.cephadm:Choosing monitor IPs and ports... 2026-03-10T10:19:41.411 INFO:tasks.cephadm:No mon roles; fabricating mons 2026-03-10T10:19:41.411 INFO:tasks.cephadm:Monitor IPs: {'mon.vm01': '192.168.123.101', 'mon.vm08': '192.168.123.108'} 2026-03-10T10:19:41.411 INFO:tasks.cephadm:Normalizing hostnames... 2026-03-10T10:19:41.411 DEBUG:teuthology.orchestra.run.vm01:> sudo hostname $(hostname -s) 2026-03-10T10:19:41.470 DEBUG:teuthology.orchestra.run.vm08:> sudo hostname $(hostname -s) 2026-03-10T10:19:41.511 INFO:tasks.cephadm:Downloading cephadm (repo https://github.com/ceph/ceph ref v17.2.0)... 2026-03-10T10:19:41.512 DEBUG:teuthology.orchestra.run.vm01:> curl --silent https://raw.githubusercontent.com/ceph/ceph/v17.2.0/src/cephadm/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-10T10:19:41.772 INFO:teuthology.orchestra.run.vm01.stdout:-rw-r--r--. 1 ubuntu ubuntu 320521 Mar 10 10:19 /home/ubuntu/cephtest/cephadm 2026-03-10T10:19:41.772 DEBUG:teuthology.orchestra.run.vm08:> curl --silent https://raw.githubusercontent.com/ceph/ceph/v17.2.0/src/cephadm/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-10T10:19:41.865 INFO:teuthology.orchestra.run.vm08.stdout:-rw-r--r--. 1 ubuntu ubuntu 320521 Mar 10 10:19 /home/ubuntu/cephtest/cephadm 2026-03-10T10:19:41.865 DEBUG:teuthology.orchestra.run.vm01:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-10T10:19:41.892 DEBUG:teuthology.orchestra.run.vm08:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-10T10:19:41.923 INFO:tasks.cephadm:Pulling image quay.io/ceph/ceph:v17.2.0 on all hosts... 2026-03-10T10:19:41.924 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 pull 2026-03-10T10:19:41.936 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 pull 2026-03-10T10:19:42.253 INFO:teuthology.orchestra.run.vm01.stderr:Pulling container image quay.io/ceph/ceph:v17.2.0... 2026-03-10T10:19:42.269 INFO:teuthology.orchestra.run.vm08.stderr:Pulling container image quay.io/ceph/ceph:v17.2.0... 2026-03-10T10:21:06.203 INFO:teuthology.orchestra.run.vm08.stdout:{ 2026-03-10T10:21:06.203 INFO:teuthology.orchestra.run.vm08.stdout: "ceph_version": "ceph version 17.2.0 (43e2e60a7559d3f46c9d53f1ca875fd499a1e35e) quincy (stable)", 2026-03-10T10:21:06.203 INFO:teuthology.orchestra.run.vm08.stdout: "image_id": "e1d6a67b021eb077ee22bf650f1a9fb1980a2cf5c36bdb9cba9eac6de8f702d9", 2026-03-10T10:21:06.203 INFO:teuthology.orchestra.run.vm08.stdout: "repo_digests": [ 2026-03-10T10:21:06.203 INFO:teuthology.orchestra.run.vm08.stdout: "quay.io/ceph/ceph@sha256:12a0a4f43413fd97a14a3d47a3451b2d2df50020835bb93db666209f3f77617a", 2026-03-10T10:21:06.203 INFO:teuthology.orchestra.run.vm08.stdout: "quay.io/ceph/ceph@sha256:cb4d698cb769b6aba05bf6ef04f41a7fe694160140347576e13bd9348514b667" 2026-03-10T10:21:06.203 INFO:teuthology.orchestra.run.vm08.stdout: ] 2026-03-10T10:21:06.203 INFO:teuthology.orchestra.run.vm08.stdout:} 2026-03-10T10:21:08.808 INFO:teuthology.orchestra.run.vm01.stdout:{ 2026-03-10T10:21:08.808 INFO:teuthology.orchestra.run.vm01.stdout: "ceph_version": "ceph version 17.2.0 (43e2e60a7559d3f46c9d53f1ca875fd499a1e35e) quincy (stable)", 2026-03-10T10:21:08.808 INFO:teuthology.orchestra.run.vm01.stdout: "image_id": "e1d6a67b021eb077ee22bf650f1a9fb1980a2cf5c36bdb9cba9eac6de8f702d9", 2026-03-10T10:21:08.808 INFO:teuthology.orchestra.run.vm01.stdout: "repo_digests": [ 2026-03-10T10:21:08.808 INFO:teuthology.orchestra.run.vm01.stdout: "quay.io/ceph/ceph@sha256:12a0a4f43413fd97a14a3d47a3451b2d2df50020835bb93db666209f3f77617a", 2026-03-10T10:21:08.808 INFO:teuthology.orchestra.run.vm01.stdout: "quay.io/ceph/ceph@sha256:cb4d698cb769b6aba05bf6ef04f41a7fe694160140347576e13bd9348514b667" 2026-03-10T10:21:08.808 INFO:teuthology.orchestra.run.vm01.stdout: ] 2026-03-10T10:21:08.808 INFO:teuthology.orchestra.run.vm01.stdout:} 2026-03-10T10:21:08.821 DEBUG:teuthology.orchestra.run.vm01:> sudo mkdir -p /etc/ceph 2026-03-10T10:21:08.847 DEBUG:teuthology.orchestra.run.vm08:> sudo mkdir -p /etc/ceph 2026-03-10T10:21:08.873 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod 777 /etc/ceph 2026-03-10T10:21:08.909 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod 777 /etc/ceph 2026-03-10T10:21:08.936 INFO:tasks.cephadm:Writing seed config... 2026-03-10T10:21:08.937 INFO:tasks.cephadm: override: [mgr] debug mgr = 20 2026-03-10T10:21:08.937 INFO:tasks.cephadm: override: [mgr] debug ms = 1 2026-03-10T10:21:08.937 INFO:tasks.cephadm: override: [mon] debug mon = 20 2026-03-10T10:21:08.937 INFO:tasks.cephadm: override: [mon] debug ms = 1 2026-03-10T10:21:08.937 INFO:tasks.cephadm: override: [mon] debug paxos = 20 2026-03-10T10:21:08.937 INFO:tasks.cephadm: override: [osd] debug ms = 1 2026-03-10T10:21:08.937 INFO:tasks.cephadm: override: [osd] debug osd = 20 2026-03-10T10:21:08.937 INFO:tasks.cephadm: override: [osd] osd mclock iops capacity threshold hdd = 49000 2026-03-10T10:21:08.937 INFO:tasks.cephadm: override: [osd] osd shutdown pgref assert = True 2026-03-10T10:21:08.937 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T10:21:08.937 DEBUG:teuthology.orchestra.run.vm01:> dd of=/home/ubuntu/cephtest/seed.ceph.conf 2026-03-10T10:21:08.963 DEBUG:tasks.cephadm:Final config: [global] # make logging friendly to teuthology log_to_file = true log_to_stderr = false log to journald = false mon cluster log to file = true mon cluster log file level = debug mon clock drift allowed = 1.000 # replicate across OSDs, not hosts osd crush chooseleaf type = 0 #osd pool default size = 2 osd pool default erasure code profile = plugin=jerasure technique=reed_sol_van k=2 m=1 crush-failure-domain=osd # enable some debugging auth debug = true ms die on old message = true ms die on bug = true debug asserts on shutdown = true # adjust warnings mon max pg per osd = 10000# >= luminous mon pg warn max object skew = 0 mon osd allow primary affinity = true mon osd allow pg remap = true mon warn on legacy crush tunables = false mon warn on crush straw calc version zero = false mon warn on no sortbitwise = false mon warn on osd down out interval zero = false mon warn on too few osds = false mon_warn_on_pool_pg_num_not_power_of_two = false # disable pg_autoscaler by default for new pools osd_pool_default_pg_autoscale_mode = off # tests delete pools mon allow pool delete = true fsid = a6b3e19a-1c6a-11f1-aaba-ff800f171c2f [osd] osd scrub load threshold = 5.0 osd scrub max interval = 600 osd mclock profile = high_recovery_ops osd recover clone overlap = true osd recovery max chunk = 1048576 osd deep scrub update digest min age = 30 osd map max advance = 10 osd memory target autotune = true # debugging osd debug shutdown = true osd debug op order = true osd debug verify stray on activate = true osd debug pg log writeout = true osd debug verify cached snaps = true osd debug verify missing on start = true osd debug misdirected ops = true osd op queue = debug_random osd op queue cut off = debug_random osd shutdown pgref assert = True bdev debug aio = true osd sloppy crc = true debug ms = 1 debug osd = 20 osd mclock iops capacity threshold hdd = 49000 [mgr] mon reweight min pgs per osd = 4 mon reweight min bytes per osd = 10 mgr/telemetry/nag = false debug mgr = 20 debug ms = 1 [mon] mon data avail warn = 5 mon mgr mkfs grace = 240 mon reweight min pgs per osd = 4 mon osd reporter subtree level = osd mon osd prime pg temp = true mon reweight min bytes per osd = 10 # rotate auth tickets quickly to exercise renewal paths auth mon ticket ttl = 660# 11m auth service ticket ttl = 240# 4m # don't complain about global id reclaim mon_warn_on_insecure_global_id_reclaim = false mon_warn_on_insecure_global_id_reclaim_allowed = false debug mon = 20 debug ms = 1 debug paxos = 20 [client.rgw] rgw cache enabled = true rgw enable ops log = true rgw enable usage log = true 2026-03-10T10:21:08.963 DEBUG:teuthology.orchestra.run.vm01:mon.vm01> sudo journalctl -f -n 0 -u ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f@mon.vm01.service 2026-03-10T10:21:09.006 INFO:tasks.cephadm:Bootstrapping... 2026-03-10T10:21:09.006 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 -v bootstrap --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f --config /home/ubuntu/cephtest/seed.ceph.conf --output-config /etc/ceph/ceph.conf --output-keyring /etc/ceph/ceph.client.admin.keyring --output-pub-ssh-key /home/ubuntu/cephtest/ceph.pub --mon-ip 192.168.123.101 --skip-admin-label && sudo chmod +r /etc/ceph/ceph.client.admin.keyring 2026-03-10T10:21:09.140 INFO:teuthology.orchestra.run.vm01.stderr:-------------------------------------------------------------------------------- 2026-03-10T10:21:09.140 INFO:teuthology.orchestra.run.vm01.stderr:cephadm ['--image', 'quay.io/ceph/ceph:v17.2.0', '-v', 'bootstrap', '--fsid', 'a6b3e19a-1c6a-11f1-aaba-ff800f171c2f', '--config', '/home/ubuntu/cephtest/seed.ceph.conf', '--output-config', '/etc/ceph/ceph.conf', '--output-keyring', '/etc/ceph/ceph.client.admin.keyring', '--output-pub-ssh-key', '/home/ubuntu/cephtest/ceph.pub', '--mon-ip', '192.168.123.101', '--skip-admin-label'] 2026-03-10T10:21:09.158 INFO:teuthology.orchestra.run.vm01.stderr:/bin/podman: 5.8.0 2026-03-10T10:21:09.161 INFO:teuthology.orchestra.run.vm01.stderr:Verifying podman|docker is present... 2026-03-10T10:21:09.176 INFO:teuthology.orchestra.run.vm01.stderr:/bin/podman: 5.8.0 2026-03-10T10:21:09.179 INFO:teuthology.orchestra.run.vm01.stderr:Verifying lvm2 is present... 2026-03-10T10:21:09.179 INFO:teuthology.orchestra.run.vm01.stderr:Verifying time synchronization is in place... 2026-03-10T10:21:09.185 INFO:teuthology.orchestra.run.vm01.stderr:systemctl: Failed to get unit file state for chrony.service: No such file or directory 2026-03-10T10:21:09.191 INFO:teuthology.orchestra.run.vm01.stderr:systemctl: inactive 2026-03-10T10:21:09.197 INFO:teuthology.orchestra.run.vm01.stderr:systemctl: enabled 2026-03-10T10:21:09.202 INFO:teuthology.orchestra.run.vm01.stderr:systemctl: active 2026-03-10T10:21:09.203 INFO:teuthology.orchestra.run.vm01.stderr:Unit chronyd.service is enabled and running 2026-03-10T10:21:09.203 INFO:teuthology.orchestra.run.vm01.stderr:Repeating the final host check... 2026-03-10T10:21:09.217 INFO:teuthology.orchestra.run.vm01.stderr:/bin/podman: 5.8.0 2026-03-10T10:21:09.220 INFO:teuthology.orchestra.run.vm01.stderr:podman (/bin/podman) version 5.8.0 is present 2026-03-10T10:21:09.220 INFO:teuthology.orchestra.run.vm01.stderr:systemctl is present 2026-03-10T10:21:09.220 INFO:teuthology.orchestra.run.vm01.stderr:lvcreate is present 2026-03-10T10:21:09.225 INFO:teuthology.orchestra.run.vm01.stderr:systemctl: Failed to get unit file state for chrony.service: No such file or directory 2026-03-10T10:21:09.231 INFO:teuthology.orchestra.run.vm01.stderr:systemctl: inactive 2026-03-10T10:21:09.236 INFO:teuthology.orchestra.run.vm01.stderr:systemctl: enabled 2026-03-10T10:21:09.240 INFO:teuthology.orchestra.run.vm01.stderr:systemctl: active 2026-03-10T10:21:09.241 INFO:teuthology.orchestra.run.vm01.stderr:Unit chronyd.service is enabled and running 2026-03-10T10:21:09.241 INFO:teuthology.orchestra.run.vm01.stderr:Host looks OK 2026-03-10T10:21:09.241 INFO:teuthology.orchestra.run.vm01.stderr:Cluster fsid: a6b3e19a-1c6a-11f1-aaba-ff800f171c2f 2026-03-10T10:21:09.241 INFO:teuthology.orchestra.run.vm01.stderr:Acquiring lock 140096076439408 on /run/cephadm/a6b3e19a-1c6a-11f1-aaba-ff800f171c2f.lock 2026-03-10T10:21:09.241 INFO:teuthology.orchestra.run.vm01.stderr:Lock 140096076439408 acquired on /run/cephadm/a6b3e19a-1c6a-11f1-aaba-ff800f171c2f.lock 2026-03-10T10:21:09.241 INFO:teuthology.orchestra.run.vm01.stderr:Verifying IP 192.168.123.101 port 3300 ... 2026-03-10T10:21:09.242 INFO:teuthology.orchestra.run.vm01.stderr:Verifying IP 192.168.123.101 port 6789 ... 2026-03-10T10:21:09.242 INFO:teuthology.orchestra.run.vm01.stderr:Base mon IP is 192.168.123.101, final addrv is [v2:192.168.123.101:3300,v1:192.168.123.101:6789] 2026-03-10T10:21:09.244 INFO:teuthology.orchestra.run.vm01.stderr:/sbin/ip: default via 192.168.123.1 dev eth0 proto dhcp src 192.168.123.101 metric 100 2026-03-10T10:21:09.244 INFO:teuthology.orchestra.run.vm01.stderr:/sbin/ip: 192.168.123.0/24 dev eth0 proto kernel scope link src 192.168.123.101 metric 100 2026-03-10T10:21:09.247 INFO:teuthology.orchestra.run.vm01.stderr:/sbin/ip: ::1 dev lo proto kernel metric 256 pref medium 2026-03-10T10:21:09.247 INFO:teuthology.orchestra.run.vm01.stderr:/sbin/ip: fe80::/64 dev eth0 proto kernel metric 1024 pref medium 2026-03-10T10:21:09.249 INFO:teuthology.orchestra.run.vm01.stderr:/sbin/ip: 1: lo: mtu 65536 state UNKNOWN qlen 1000 2026-03-10T10:21:09.249 INFO:teuthology.orchestra.run.vm01.stderr:/sbin/ip: inet6 ::1/128 scope host 2026-03-10T10:21:09.249 INFO:teuthology.orchestra.run.vm01.stderr:/sbin/ip: valid_lft forever preferred_lft forever 2026-03-10T10:21:09.249 INFO:teuthology.orchestra.run.vm01.stderr:/sbin/ip: 2: eth0: mtu 1500 state UP qlen 1000 2026-03-10T10:21:09.249 INFO:teuthology.orchestra.run.vm01.stderr:/sbin/ip: inet6 fe80::5055:ff:fe00:1/64 scope link noprefixroute 2026-03-10T10:21:09.249 INFO:teuthology.orchestra.run.vm01.stderr:/sbin/ip: valid_lft forever preferred_lft forever 2026-03-10T10:21:09.250 INFO:teuthology.orchestra.run.vm01.stderr:Mon IP `192.168.123.101` is in CIDR network `192.168.123.0/24` 2026-03-10T10:21:09.250 INFO:teuthology.orchestra.run.vm01.stderr:- internal network (--cluster-network) has not been provided, OSD replication will default to the public_network 2026-03-10T10:21:09.250 INFO:teuthology.orchestra.run.vm01.stderr:Pulling container image quay.io/ceph/ceph:v17.2.0... 2026-03-10T10:21:09.268 INFO:teuthology.orchestra.run.vm01.stderr:/bin/podman: Trying to pull quay.io/ceph/ceph:v17.2.0... 2026-03-10T10:21:10.638 INFO:teuthology.orchestra.run.vm01.stderr:/bin/podman: Getting image source signatures 2026-03-10T10:21:10.638 INFO:teuthology.orchestra.run.vm01.stderr:/bin/podman: Copying blob sha256:33ca8fff7868c4dc0c11e09bca97c720eb9cfbab7221216754367dd8de70388a 2026-03-10T10:21:10.638 INFO:teuthology.orchestra.run.vm01.stderr:/bin/podman: Copying blob sha256:89b4a75bc2d8500f15463747507c9623df43886c134463e7f0527e70900e7a7b 2026-03-10T10:21:10.638 INFO:teuthology.orchestra.run.vm01.stderr:/bin/podman: Copying blob sha256:c32ab78b488d0c72f64eded765c0cf6b5bf2c75dab66cb62a9d367fa6ec42513 2026-03-10T10:21:10.638 INFO:teuthology.orchestra.run.vm01.stderr:/bin/podman: Copying blob sha256:a70843738bb77e1ab9c1f85969ebdfa55f178e746be081d1cb4f94011f69eb7c 2026-03-10T10:21:10.638 INFO:teuthology.orchestra.run.vm01.stderr:/bin/podman: Copying blob sha256:599d07cb321ff0a3c82224e1138fc685793fa69b93ed5780415751a5f7e4b8c2 2026-03-10T10:21:10.639 INFO:teuthology.orchestra.run.vm01.stderr:/bin/podman: Copying config sha256:e1d6a67b021eb077ee22bf650f1a9fb1980a2cf5c36bdb9cba9eac6de8f702d9 2026-03-10T10:21:10.641 INFO:teuthology.orchestra.run.vm01.stderr:/bin/podman: Writing manifest to image destination 2026-03-10T10:21:10.646 INFO:teuthology.orchestra.run.vm01.stderr:/bin/podman: e1d6a67b021eb077ee22bf650f1a9fb1980a2cf5c36bdb9cba9eac6de8f702d9 2026-03-10T10:21:10.786 INFO:teuthology.orchestra.run.vm01.stderr:ceph: ceph version 17.2.0 (43e2e60a7559d3f46c9d53f1ca875fd499a1e35e) quincy (stable) 2026-03-10T10:21:10.816 INFO:teuthology.orchestra.run.vm01.stderr:Ceph version: ceph version 17.2.0 (43e2e60a7559d3f46c9d53f1ca875fd499a1e35e) quincy (stable) 2026-03-10T10:21:10.816 INFO:teuthology.orchestra.run.vm01.stderr:Extracting ceph user uid/gid from container image... 2026-03-10T10:21:10.923 INFO:teuthology.orchestra.run.vm01.stderr:stat: 167 167 2026-03-10T10:21:10.940 INFO:teuthology.orchestra.run.vm01.stderr:Creating initial keys... 2026-03-10T10:21:11.012 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph-authtool: AQCX8K9psfGgABAAJ1EEleVdC7pFu2QzOiVGNg== 2026-03-10T10:21:11.111 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph-authtool: AQCX8K9pb9OKBhAAupgpYdvPoOauZgW6qIEPtA== 2026-03-10T10:21:11.199 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph-authtool: AQCX8K9psrLPCxAAWE72IcabeoaqeaelbjZL/g== 2026-03-10T10:21:11.239 INFO:teuthology.orchestra.run.vm01.stderr:Creating initial monmap... 2026-03-10T10:21:11.464 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/monmaptool: /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-10T10:21:11.464 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/monmaptool: setting min_mon_release = octopus 2026-03-10T10:21:11.464 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/monmaptool: /usr/bin/monmaptool: set fsid to a6b3e19a-1c6a-11f1-aaba-ff800f171c2f 2026-03-10T10:21:11.464 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/monmaptool: /usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-10T10:21:11.484 INFO:teuthology.orchestra.run.vm01.stderr:monmaptool for vm01 [v2:192.168.123.101:3300,v1:192.168.123.101:6789] on /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-10T10:21:11.485 INFO:teuthology.orchestra.run.vm01.stderr:setting min_mon_release = octopus 2026-03-10T10:21:11.485 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/monmaptool: set fsid to a6b3e19a-1c6a-11f1-aaba-ff800f171c2f 2026-03-10T10:21:11.485 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-10T10:21:11.485 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-10T10:21:11.485 INFO:teuthology.orchestra.run.vm01.stderr:Creating mon... 2026-03-10T10:21:11.627 INFO:teuthology.orchestra.run.vm01.stderr:create mon.vm01 on 2026-03-10T10:21:11.781 INFO:teuthology.orchestra.run.vm01.stderr:systemctl: Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /etc/systemd/system/ceph.target. 2026-03-10T10:21:11.910 INFO:teuthology.orchestra.run.vm01.stderr:systemctl: Created symlink /etc/systemd/system/multi-user.target.wants/ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f.target → /etc/systemd/system/ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f.target. 2026-03-10T10:21:11.910 INFO:teuthology.orchestra.run.vm01.stderr:systemctl: Created symlink /etc/systemd/system/ceph.target.wants/ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f.target → /etc/systemd/system/ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f.target. 2026-03-10T10:21:12.180 INFO:teuthology.orchestra.run.vm01.stderr:systemctl: Failed to reset failed state of unit ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f@mon.vm01.service: Unit ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f@mon.vm01.service not loaded. 2026-03-10T10:21:12.187 INFO:teuthology.orchestra.run.vm01.stderr:systemctl: Created symlink /etc/systemd/system/ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f.target.wants/ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f@mon.vm01.service → /etc/systemd/system/ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f@.service. 2026-03-10T10:21:12.635 INFO:teuthology.orchestra.run.vm01.stderr:firewalld does not appear to be present 2026-03-10T10:21:12.635 INFO:teuthology.orchestra.run.vm01.stderr:Not possible to enable service . firewalld.service is not available 2026-03-10T10:21:12.635 INFO:teuthology.orchestra.run.vm01.stderr:Waiting for mon to start... 2026-03-10T10:21:12.635 INFO:teuthology.orchestra.run.vm01.stderr:Waiting for mon... 2026-03-10T10:21:12.837 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: cluster: 2026-03-10T10:21:12.837 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: id: a6b3e19a-1c6a-11f1-aaba-ff800f171c2f 2026-03-10T10:21:12.837 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: health: HEALTH_OK 2026-03-10T10:21:12.837 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: 2026-03-10T10:21:12.837 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: services: 2026-03-10T10:21:12.837 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: mon: 1 daemons, quorum vm01 (age 0.172823s) 2026-03-10T10:21:12.837 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: mgr: no daemons active 2026-03-10T10:21:12.837 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: osd: 0 osds: 0 up, 0 in 2026-03-10T10:21:12.838 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: 2026-03-10T10:21:12.838 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: data: 2026-03-10T10:21:12.838 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: pools: 0 pools, 0 pgs 2026-03-10T10:21:12.838 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: objects: 0 objects, 0 B 2026-03-10T10:21:12.838 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: usage: 0 B used, 0 B / 0 B avail 2026-03-10T10:21:12.838 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: pgs: 2026-03-10T10:21:12.838 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: 2026-03-10T10:21:12.887 INFO:teuthology.orchestra.run.vm01.stderr:mon is available 2026-03-10T10:21:12.887 INFO:teuthology.orchestra.run.vm01.stderr:Assimilating anything we can from ceph.conf... 2026-03-10T10:21:13.129 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: 2026-03-10T10:21:13.129 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: [global] 2026-03-10T10:21:13.129 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: fsid = a6b3e19a-1c6a-11f1-aaba-ff800f171c2f 2026-03-10T10:21:13.130 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: mon_host = [v2:192.168.123.101:3300,v1:192.168.123.101:6789] 2026-03-10T10:21:13.130 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: mon_osd_allow_pg_remap = true 2026-03-10T10:21:13.130 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: mon_osd_allow_primary_affinity = true 2026-03-10T10:21:13.130 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: mon_warn_on_no_sortbitwise = false 2026-03-10T10:21:13.130 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: osd_crush_chooseleaf_type = 0 2026-03-10T10:21:13.130 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: 2026-03-10T10:21:13.130 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: [mgr] 2026-03-10T10:21:13.130 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: mgr/telemetry/nag = false 2026-03-10T10:21:13.130 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: 2026-03-10T10:21:13.130 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: [osd] 2026-03-10T10:21:13.130 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: osd_map_max_advance = 10 2026-03-10T10:21:13.130 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: osd_mclock_iops_capacity_threshold_hdd = 49000 2026-03-10T10:21:13.130 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: osd_sloppy_crc = true 2026-03-10T10:21:13.192 INFO:teuthology.orchestra.run.vm01.stderr:Generating new minimal ceph.conf... 2026-03-10T10:21:13.413 INFO:teuthology.orchestra.run.vm01.stderr:Restarting the monitor... 2026-03-10T10:21:13.691 INFO:teuthology.orchestra.run.vm01.stderr:Setting mon public_network to 192.168.123.0/24 2026-03-10T10:21:14.074 INFO:teuthology.orchestra.run.vm01.stderr:Wrote config to /etc/ceph/ceph.conf 2026-03-10T10:21:14.074 INFO:teuthology.orchestra.run.vm01.stderr:Wrote keyring to /etc/ceph/ceph.client.admin.keyring 2026-03-10T10:21:14.074 INFO:teuthology.orchestra.run.vm01.stderr:Creating mgr... 2026-03-10T10:21:14.074 INFO:teuthology.orchestra.run.vm01.stderr:Verifying port 9283 ... 2026-03-10T10:21:14.214 INFO:teuthology.orchestra.run.vm01.stderr:systemctl: Failed to reset failed state of unit ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f@mgr.vm01.nkkzmc.service: Unit ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f@mgr.vm01.nkkzmc.service not loaded. 2026-03-10T10:21:14.221 INFO:teuthology.orchestra.run.vm01.stderr:systemctl: Created symlink /etc/systemd/system/ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f.target.wants/ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f@mgr.vm01.nkkzmc.service → /etc/systemd/system/ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f@.service. 2026-03-10T10:21:14.517 INFO:teuthology.orchestra.run.vm01.stderr:firewalld does not appear to be present 2026-03-10T10:21:14.517 INFO:teuthology.orchestra.run.vm01.stderr:Not possible to enable service . firewalld.service is not available 2026-03-10T10:21:14.517 INFO:teuthology.orchestra.run.vm01.stderr:firewalld does not appear to be present 2026-03-10T10:21:14.517 INFO:teuthology.orchestra.run.vm01.stderr:Not possible to open ports <[9283]>. firewalld.service is not available 2026-03-10T10:21:14.517 INFO:teuthology.orchestra.run.vm01.stderr:Waiting for mgr to start... 2026-03-10T10:21:14.517 INFO:teuthology.orchestra.run.vm01.stderr:Waiting for mgr... 2026-03-10T10:21:14.787 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: 2026-03-10T10:21:14.787 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: { 2026-03-10T10:21:14.787 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "fsid": "a6b3e19a-1c6a-11f1-aaba-ff800f171c2f", 2026-03-10T10:21:14.787 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "health": { 2026-03-10T10:21:14.787 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "status": "HEALTH_OK", 2026-03-10T10:21:14.787 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "checks": {}, 2026-03-10T10:21:14.787 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "mutes": [] 2026-03-10T10:21:14.787 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: }, 2026-03-10T10:21:14.787 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "election_epoch": 5, 2026-03-10T10:21:14.787 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "quorum": [ 2026-03-10T10:21:14.787 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: 0 2026-03-10T10:21:14.787 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: ], 2026-03-10T10:21:14.787 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "quorum_names": [ 2026-03-10T10:21:14.787 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "vm01" 2026-03-10T10:21:14.788 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: ], 2026-03-10T10:21:14.788 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "quorum_age": 1, 2026-03-10T10:21:14.788 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "monmap": { 2026-03-10T10:21:14.788 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-10T10:21:14.788 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "min_mon_release_name": "quincy", 2026-03-10T10:21:14.788 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_mons": 1 2026-03-10T10:21:14.788 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: }, 2026-03-10T10:21:14.788 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "osdmap": { 2026-03-10T10:21:14.788 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-10T10:21:14.788 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_osds": 0, 2026-03-10T10:21:14.788 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_up_osds": 0, 2026-03-10T10:21:14.788 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "osd_up_since": 0, 2026-03-10T10:21:14.788 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_in_osds": 0, 2026-03-10T10:21:14.788 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "osd_in_since": 0, 2026-03-10T10:21:14.788 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_remapped_pgs": 0 2026-03-10T10:21:14.788 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: }, 2026-03-10T10:21:14.788 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "pgmap": { 2026-03-10T10:21:14.788 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "pgs_by_state": [], 2026-03-10T10:21:14.788 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_pgs": 0, 2026-03-10T10:21:14.788 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_pools": 0, 2026-03-10T10:21:14.788 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_objects": 0, 2026-03-10T10:21:14.788 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "data_bytes": 0, 2026-03-10T10:21:14.788 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "bytes_used": 0, 2026-03-10T10:21:14.788 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "bytes_avail": 0, 2026-03-10T10:21:14.788 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "bytes_total": 0 2026-03-10T10:21:14.788 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: }, 2026-03-10T10:21:14.789 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "fsmap": { 2026-03-10T10:21:14.789 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-10T10:21:14.789 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "by_rank": [], 2026-03-10T10:21:14.789 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "up:standby": 0 2026-03-10T10:21:14.789 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: }, 2026-03-10T10:21:14.789 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "mgrmap": { 2026-03-10T10:21:14.789 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "available": false, 2026-03-10T10:21:14.789 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_standbys": 0, 2026-03-10T10:21:14.789 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "modules": [ 2026-03-10T10:21:14.789 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "iostat", 2026-03-10T10:21:14.789 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "nfs", 2026-03-10T10:21:14.789 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "restful" 2026-03-10T10:21:14.789 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: ], 2026-03-10T10:21:14.789 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "services": {} 2026-03-10T10:21:14.789 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: }, 2026-03-10T10:21:14.789 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "servicemap": { 2026-03-10T10:21:14.789 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-10T10:21:14.789 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "modified": "2026-03-10T10:21:12.663524+0000", 2026-03-10T10:21:14.789 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "services": {} 2026-03-10T10:21:14.789 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: }, 2026-03-10T10:21:14.789 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "progress_events": {} 2026-03-10T10:21:14.789 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: } 2026-03-10T10:21:14.819 INFO:teuthology.orchestra.run.vm01.stderr:mgr not available, waiting (1/15)... 2026-03-10T10:21:17.077 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: 2026-03-10T10:21:17.077 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: { 2026-03-10T10:21:17.077 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "fsid": "a6b3e19a-1c6a-11f1-aaba-ff800f171c2f", 2026-03-10T10:21:17.077 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "health": { 2026-03-10T10:21:17.077 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "status": "HEALTH_OK", 2026-03-10T10:21:17.077 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "checks": {}, 2026-03-10T10:21:17.077 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "mutes": [] 2026-03-10T10:21:17.077 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: }, 2026-03-10T10:21:17.077 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "election_epoch": 5, 2026-03-10T10:21:17.077 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "quorum": [ 2026-03-10T10:21:17.077 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: 0 2026-03-10T10:21:17.077 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: ], 2026-03-10T10:21:17.077 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "quorum_names": [ 2026-03-10T10:21:17.077 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "vm01" 2026-03-10T10:21:17.077 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: ], 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "quorum_age": 3, 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "monmap": { 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "min_mon_release_name": "quincy", 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_mons": 1 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: }, 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "osdmap": { 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_osds": 0, 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_up_osds": 0, 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "osd_up_since": 0, 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_in_osds": 0, 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "osd_in_since": 0, 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_remapped_pgs": 0 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: }, 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "pgmap": { 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "pgs_by_state": [], 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_pgs": 0, 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_pools": 0, 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_objects": 0, 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "data_bytes": 0, 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "bytes_used": 0, 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "bytes_avail": 0, 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "bytes_total": 0 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: }, 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "fsmap": { 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "by_rank": [], 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "up:standby": 0 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: }, 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "mgrmap": { 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "available": false, 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_standbys": 0, 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "modules": [ 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "iostat", 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "nfs", 2026-03-10T10:21:17.078 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "restful" 2026-03-10T10:21:17.079 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: ], 2026-03-10T10:21:17.079 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "services": {} 2026-03-10T10:21:17.079 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: }, 2026-03-10T10:21:17.079 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "servicemap": { 2026-03-10T10:21:17.079 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-10T10:21:17.079 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "modified": "2026-03-10T10:21:12.663524+0000", 2026-03-10T10:21:17.079 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "services": {} 2026-03-10T10:21:17.079 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: }, 2026-03-10T10:21:17.079 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "progress_events": {} 2026-03-10T10:21:17.079 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: } 2026-03-10T10:21:17.109 INFO:teuthology.orchestra.run.vm01.stderr:mgr not available, waiting (2/15)... 2026-03-10T10:21:19.291 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: 2026-03-10T10:21:19.291 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: { 2026-03-10T10:21:19.291 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "fsid": "a6b3e19a-1c6a-11f1-aaba-ff800f171c2f", 2026-03-10T10:21:19.291 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "health": { 2026-03-10T10:21:19.291 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "status": "HEALTH_OK", 2026-03-10T10:21:19.291 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "checks": {}, 2026-03-10T10:21:19.291 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "mutes": [] 2026-03-10T10:21:19.291 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: }, 2026-03-10T10:21:19.291 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "election_epoch": 5, 2026-03-10T10:21:19.291 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "quorum": [ 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: 0 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: ], 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "quorum_names": [ 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "vm01" 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: ], 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "quorum_age": 5, 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "monmap": { 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "min_mon_release_name": "quincy", 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_mons": 1 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: }, 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "osdmap": { 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_osds": 0, 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_up_osds": 0, 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "osd_up_since": 0, 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_in_osds": 0, 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "osd_in_since": 0, 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_remapped_pgs": 0 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: }, 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "pgmap": { 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "pgs_by_state": [], 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_pgs": 0, 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_pools": 0, 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_objects": 0, 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "data_bytes": 0, 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "bytes_used": 0, 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "bytes_avail": 0, 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "bytes_total": 0 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: }, 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "fsmap": { 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "by_rank": [], 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "up:standby": 0 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: }, 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "mgrmap": { 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "available": false, 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_standbys": 0, 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "modules": [ 2026-03-10T10:21:19.292 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "iostat", 2026-03-10T10:21:19.293 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "nfs", 2026-03-10T10:21:19.293 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "restful" 2026-03-10T10:21:19.293 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: ], 2026-03-10T10:21:19.293 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "services": {} 2026-03-10T10:21:19.293 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: }, 2026-03-10T10:21:19.293 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "servicemap": { 2026-03-10T10:21:19.293 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-10T10:21:19.293 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "modified": "2026-03-10T10:21:12.663524+0000", 2026-03-10T10:21:19.293 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "services": {} 2026-03-10T10:21:19.293 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: }, 2026-03-10T10:21:19.293 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "progress_events": {} 2026-03-10T10:21:19.293 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: } 2026-03-10T10:21:19.321 INFO:teuthology.orchestra.run.vm01.stderr:mgr not available, waiting (3/15)... 2026-03-10T10:21:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:19 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/3534277774' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T10:21:21.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:20 vm01 ceph-mon[48597]: mgrmap e3: vm01.nkkzmc(active, since 1.0726s) 2026-03-10T10:21:21.739 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: 2026-03-10T10:21:21.739 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: { 2026-03-10T10:21:21.739 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "fsid": "a6b3e19a-1c6a-11f1-aaba-ff800f171c2f", 2026-03-10T10:21:21.739 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "health": { 2026-03-10T10:21:21.739 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "status": "HEALTH_OK", 2026-03-10T10:21:21.739 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "checks": {}, 2026-03-10T10:21:21.739 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "mutes": [] 2026-03-10T10:21:21.739 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: }, 2026-03-10T10:21:21.739 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "election_epoch": 5, 2026-03-10T10:21:21.739 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "quorum": [ 2026-03-10T10:21:21.739 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: 0 2026-03-10T10:21:21.739 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: ], 2026-03-10T10:21:21.740 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "quorum_names": [ 2026-03-10T10:21:21.740 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "vm01" 2026-03-10T10:21:21.740 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: ], 2026-03-10T10:21:21.740 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "quorum_age": 8, 2026-03-10T10:21:21.740 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "monmap": { 2026-03-10T10:21:21.740 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-10T10:21:21.740 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "min_mon_release_name": "quincy", 2026-03-10T10:21:21.740 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_mons": 1 2026-03-10T10:21:21.740 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: }, 2026-03-10T10:21:21.740 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "osdmap": { 2026-03-10T10:21:21.740 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-10T10:21:21.740 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_osds": 0, 2026-03-10T10:21:21.740 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_up_osds": 0, 2026-03-10T10:21:21.740 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "osd_up_since": 0, 2026-03-10T10:21:21.740 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_in_osds": 0, 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "osd_in_since": 0, 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_remapped_pgs": 0 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: }, 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "pgmap": { 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "pgs_by_state": [], 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_pgs": 0, 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_pools": 0, 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_objects": 0, 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "data_bytes": 0, 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "bytes_used": 0, 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "bytes_avail": 0, 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "bytes_total": 0 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: }, 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "fsmap": { 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "by_rank": [], 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "up:standby": 0 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: }, 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "mgrmap": { 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "available": true, 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_standbys": 0, 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "modules": [ 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "iostat", 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "nfs", 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "restful" 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: ], 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "services": {} 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: }, 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "servicemap": { 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "modified": "2026-03-10T10:21:12.663524+0000", 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "services": {} 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: }, 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "progress_events": {} 2026-03-10T10:21:21.741 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: } 2026-03-10T10:21:21.788 INFO:teuthology.orchestra.run.vm01.stderr:mgr is available 2026-03-10T10:21:22.015 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:21 vm01 ceph-mon[48597]: mgrmap e4: vm01.nkkzmc(active, since 2s) 2026-03-10T10:21:22.015 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:21 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/4128491879' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T10:21:22.016 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: 2026-03-10T10:21:22.016 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: [global] 2026-03-10T10:21:22.016 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: fsid = a6b3e19a-1c6a-11f1-aaba-ff800f171c2f 2026-03-10T10:21:22.016 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: mon_osd_allow_pg_remap = true 2026-03-10T10:21:22.016 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: mon_osd_allow_primary_affinity = true 2026-03-10T10:21:22.016 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: mon_warn_on_no_sortbitwise = false 2026-03-10T10:21:22.016 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: osd_crush_chooseleaf_type = 0 2026-03-10T10:21:22.016 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: 2026-03-10T10:21:22.016 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: [mgr] 2026-03-10T10:21:22.016 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: mgr/telemetry/nag = false 2026-03-10T10:21:22.016 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: 2026-03-10T10:21:22.016 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: [osd] 2026-03-10T10:21:22.016 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: osd_map_max_advance = 10 2026-03-10T10:21:22.016 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: osd_mclock_iops_capacity_threshold_hdd = 49000 2026-03-10T10:21:22.016 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: osd_sloppy_crc = true 2026-03-10T10:21:22.063 INFO:teuthology.orchestra.run.vm01.stderr:Enabling cephadm module... 2026-03-10T10:21:23.207 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:22 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/678454227' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-10T10:21:23.207 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:22 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/4017808751' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-10T10:21:23.207 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: { 2026-03-10T10:21:23.207 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "epoch": 5, 2026-03-10T10:21:23.207 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "available": true, 2026-03-10T10:21:23.207 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "active_name": "vm01.nkkzmc", 2026-03-10T10:21:23.207 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_standby": 0 2026-03-10T10:21:23.207 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: } 2026-03-10T10:21:23.239 INFO:teuthology.orchestra.run.vm01.stderr:Waiting for the mgr to restart... 2026-03-10T10:21:23.239 INFO:teuthology.orchestra.run.vm01.stderr:Waiting for mgr epoch 5... 2026-03-10T10:21:24.241 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:23 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/4017808751' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-10T10:21:24.241 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:23 vm01 ceph-mon[48597]: mgrmap e5: vm01.nkkzmc(active, since 4s) 2026-03-10T10:21:24.241 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:23 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/1867654608' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T10:21:27.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:27 vm01 ceph-mon[48597]: Active manager daemon vm01.nkkzmc restarted 2026-03-10T10:21:27.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:27 vm01 ceph-mon[48597]: Activating manager daemon vm01.nkkzmc 2026-03-10T10:21:27.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:27 vm01 ceph-mon[48597]: osdmap e2: 0 total, 0 up, 0 in 2026-03-10T10:21:28.243 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: { 2026-03-10T10:21:28.244 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "mgrmap_epoch": 7, 2026-03-10T10:21:28.244 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "initialized": true 2026-03-10T10:21:28.244 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: } 2026-03-10T10:21:28.280 INFO:teuthology.orchestra.run.vm01.stderr:mgr epoch 5 is available 2026-03-10T10:21:28.280 INFO:teuthology.orchestra.run.vm01.stderr:Setting orchestrator backend to cephadm... 2026-03-10T10:21:28.483 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:28 vm01 ceph-mon[48597]: mgrmap e6: vm01.nkkzmc(active, starting, since 0.0540488s) 2026-03-10T10:21:28.483 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:28 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-10T10:21:28.484 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:28 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mgr metadata", "who": "vm01.nkkzmc", "id": "vm01.nkkzmc"}]: dispatch 2026-03-10T10:21:28.484 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:28 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T10:21:28.484 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:28 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T10:21:28.484 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:28 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T10:21:28.484 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:28 vm01 ceph-mon[48597]: Manager daemon vm01.nkkzmc is now available 2026-03-10T10:21:28.484 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:28 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:28.484 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:28 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:28.484 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:28 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:21:28.484 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:28 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:21:28.484 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:28 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:21:28.484 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:28 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:21:28.484 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:28 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:21:28.818 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: value unchanged 2026-03-10T10:21:28.864 INFO:teuthology.orchestra.run.vm01.stderr:Generating ssh key... 2026-03-10T10:21:29.422 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:29 vm01 ceph-mon[48597]: mgrmap e7: vm01.nkkzmc(active, since 1.06247s) 2026-03-10T10:21:29.423 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:29 vm01 ceph-mon[48597]: [10/Mar/2026:10:21:28] ENGINE Bus STARTING 2026-03-10T10:21:29.423 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:29 vm01 ceph-mon[48597]: [10/Mar/2026:10:21:28] ENGINE Serving on https://192.168.123.101:7150 2026-03-10T10:21:29.423 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:29 vm01 ceph-mon[48597]: [10/Mar/2026:10:21:28] ENGINE Bus STARTED 2026-03-10T10:21:29.423 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:29 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:29.423 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:29 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:21:29.423 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:29 vm01 ceph-mon[48597]: from='client.14132 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:21:29.423 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:29 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:29.423 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:29 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:21:29.423 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:29 vm01 ceph-mon[48597]: from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:21:29.423 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:29 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:29.423 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:29 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:29.634 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDRXCUjFz3/Mho9h06Q/GCm4LjUEOOOe1LThXg5SOOwvJ6pBA0rhao9Bpq83CsmV47txnlcFQaalL9kjvV7HhDMtU5+nKti1O61CoSaYbOU4DBMGDrkEzpF+56LCQjJHL9CCtEvPjlIDuogGrN3mg8Xplss7Wb0+ckT5hqMST+vguEFhUrkpr3a3GVAGqGY/tgUDrufgNc96P3GPKpUDZesAA+AMvbhoFF++mnMly9NR8TlLHEKT0k96FJDM4XdudWc2R82//PTJPhCT+npxn7qUaBTDkmYL/R+RlpNuNfsdyxDcJso6EzKoM4wKJASr8UBS6LhCt680X6C41Gm/DLhZqG1Rh8oaQrq7nAb0zh3OxlVYGCgvuBl1RVMyq9U5sQJMiLS1AxrFSKvociO6oNsLFunV3VCQUKdm0UeGrs0aSP8l/2s0a/qxxtKBvrjgtPNqtONPBlpxbCJIyT6tLivbLvZT32NFYRuJN8kvFYB5gyQYRjMENVXO4NHoBeoa5c= ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f 2026-03-10T10:21:29.761 INFO:teuthology.orchestra.run.vm01.stderr:Wrote public SSH key to /home/ubuntu/cephtest/ceph.pub 2026-03-10T10:21:29.761 INFO:teuthology.orchestra.run.vm01.stderr:Adding key to root@localhost authorized_keys... 2026-03-10T10:21:29.761 INFO:teuthology.orchestra.run.vm01.stderr:Adding host vm01... 2026-03-10T10:21:30.712 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: Added host 'vm01' with addr '192.168.123.101' 2026-03-10T10:21:30.749 INFO:teuthology.orchestra.run.vm01.stderr:Deploying mon service with default placement... 2026-03-10T10:21:31.035 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:30 vm01 ceph-mon[48597]: from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:21:31.035 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:30 vm01 ceph-mon[48597]: Generating ssh key... 2026-03-10T10:21:31.035 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:30 vm01 ceph-mon[48597]: from='client.14138 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:21:31.035 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:30 vm01 ceph-mon[48597]: mgrmap e8: vm01.nkkzmc(active, since 2s) 2026-03-10T10:21:31.035 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:30 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:31.036 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:30 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:21:31.036 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: Scheduled mon update... 2026-03-10T10:21:31.070 INFO:teuthology.orchestra.run.vm01.stderr:Deploying mgr service with default placement... 2026-03-10T10:21:31.315 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: Scheduled mgr update... 2026-03-10T10:21:31.347 INFO:teuthology.orchestra.run.vm01.stderr:Deploying crash service with default placement... 2026-03-10T10:21:31.594 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: Scheduled crash update... 2026-03-10T10:21:31.640 INFO:teuthology.orchestra.run.vm01.stderr:Deploying prometheus service with default placement... 2026-03-10T10:21:31.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:31 vm01 ceph-mon[48597]: from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm01", "addr": "192.168.123.101", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:21:31.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:31 vm01 ceph-mon[48597]: Deploying cephadm binary to vm01 2026-03-10T10:21:31.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:31 vm01 ceph-mon[48597]: Added host vm01 2026-03-10T10:21:31.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:31 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:31.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:31 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:31.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:31 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:31.905 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: Scheduled prometheus update... 2026-03-10T10:21:31.955 INFO:teuthology.orchestra.run.vm01.stderr:Deploying grafana service with default placement... 2026-03-10T10:21:32.236 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: Scheduled grafana update... 2026-03-10T10:21:32.306 INFO:teuthology.orchestra.run.vm01.stderr:Deploying node-exporter service with default placement... 2026-03-10T10:21:32.617 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: Scheduled node-exporter update... 2026-03-10T10:21:32.653 INFO:teuthology.orchestra.run.vm01.stderr:Deploying alertmanager service with default placement... 2026-03-10T10:21:33.583 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: Scheduled alertmanager update... 2026-03-10T10:21:33.805 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:33 vm01 ceph-mon[48597]: from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:21:33.805 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:33 vm01 ceph-mon[48597]: Saving service mon spec with placement count:5 2026-03-10T10:21:33.805 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:33 vm01 ceph-mon[48597]: from='client.14144 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:21:33.805 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:33 vm01 ceph-mon[48597]: Saving service mgr spec with placement count:2 2026-03-10T10:21:33.805 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:33 vm01 ceph-mon[48597]: from='client.14146 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "crash", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:21:33.806 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:33 vm01 ceph-mon[48597]: Saving service crash spec with placement * 2026-03-10T10:21:33.806 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:33 vm01 ceph-mon[48597]: from='client.14148 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "prometheus", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:21:33.806 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:33 vm01 ceph-mon[48597]: Saving service prometheus spec with placement count:1 2026-03-10T10:21:33.806 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:33 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:33.806 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:33 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:33.806 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:33 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:33.806 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:33 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:33.806 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:33 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:34.482 INFO:teuthology.orchestra.run.vm01.stderr:Enabling the dashboard module... 2026-03-10T10:21:34.729 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:34 vm01 ceph-mon[48597]: from='client.14150 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "grafana", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:21:34.729 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:34 vm01 ceph-mon[48597]: Saving service grafana spec with placement count:1 2026-03-10T10:21:34.729 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:34 vm01 ceph-mon[48597]: from='client.14152 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "node-exporter", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:21:34.729 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:34 vm01 ceph-mon[48597]: Saving service node-exporter spec with placement * 2026-03-10T10:21:34.729 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:34 vm01 ceph-mon[48597]: from='client.14154 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "alertmanager", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:21:34.729 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:34 vm01 ceph-mon[48597]: Saving service alertmanager spec with placement count:1 2026-03-10T10:21:34.729 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:34 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:34.729 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:34 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:34.729 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:34 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/1320355263' entity='client.admin' 2026-03-10T10:21:34.729 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:34 vm01 ceph-mon[48597]: from='mgr.14120 192.168.123.101:0/1336742168' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:34.729 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:34 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/568719367' entity='client.admin' 2026-03-10T10:21:35.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:35 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/747270222' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-10T10:21:35.877 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: { 2026-03-10T10:21:35.877 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "epoch": 9, 2026-03-10T10:21:35.877 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "available": true, 2026-03-10T10:21:35.877 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "active_name": "vm01.nkkzmc", 2026-03-10T10:21:35.877 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "num_standby": 0 2026-03-10T10:21:35.877 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: } 2026-03-10T10:21:35.911 INFO:teuthology.orchestra.run.vm01.stderr:Waiting for the mgr to restart... 2026-03-10T10:21:35.911 INFO:teuthology.orchestra.run.vm01.stderr:Waiting for mgr epoch 9... 2026-03-10T10:21:36.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:36 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/747270222' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-10T10:21:36.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:36 vm01 ceph-mon[48597]: mgrmap e9: vm01.nkkzmc(active, since 8s) 2026-03-10T10:21:36.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:36 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/1629898463' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T10:21:40.615 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:40 vm01 ceph-mon[48597]: Active manager daemon vm01.nkkzmc restarted 2026-03-10T10:21:40.615 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:40 vm01 ceph-mon[48597]: Activating manager daemon vm01.nkkzmc 2026-03-10T10:21:40.615 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:40 vm01 ceph-mon[48597]: osdmap e3: 0 total, 0 up, 0 in 2026-03-10T10:21:41.201 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: { 2026-03-10T10:21:41.202 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "mgrmap_epoch": 11, 2026-03-10T10:21:41.202 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: "initialized": true 2026-03-10T10:21:41.202 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: } 2026-03-10T10:21:41.253 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:41 vm01 ceph-mon[48597]: mgrmap e10: vm01.nkkzmc(active, starting, since 0.0554643s) 2026-03-10T10:21:41.253 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:41 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-10T10:21:41.253 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:41 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mgr metadata", "who": "vm01.nkkzmc", "id": "vm01.nkkzmc"}]: dispatch 2026-03-10T10:21:41.253 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:41 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T10:21:41.253 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:41 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T10:21:41.253 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:41 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T10:21:41.253 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:41 vm01 ceph-mon[48597]: Manager daemon vm01.nkkzmc is now available 2026-03-10T10:21:41.253 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:41 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:21:41.253 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:41 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:21:41.253 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:41 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:21:41.253 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:41 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:21:41.253 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:41 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:41.253 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:41 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:41.253 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:41 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:41.263 INFO:teuthology.orchestra.run.vm01.stderr:mgr epoch 9 is available 2026-03-10T10:21:41.263 INFO:teuthology.orchestra.run.vm01.stderr:Generating a dashboard self-signed certificate... 2026-03-10T10:21:41.712 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: Self-signed certificate created 2026-03-10T10:21:41.767 INFO:teuthology.orchestra.run.vm01.stderr:Creating initial admin user... 2026-03-10T10:21:42.162 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: {"username": "admin", "password": "$2b$12$STwmorViva/JI0g9Yb73KOffOUY.VX0soelhKq1uImTFsHTVNvMIO", "roles": ["administrator"], "name": null, "email": null, "lastUpdate": 1773138102, "enabled": true, "pwdExpirationDate": null, "pwdUpdateRequired": true} 2026-03-10T10:21:42.195 INFO:teuthology.orchestra.run.vm01.stderr:Fetching dashboard port number... 2026-03-10T10:21:42.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:42 vm01 ceph-mon[48597]: [10/Mar/2026:10:21:40] ENGINE Bus STARTING 2026-03-10T10:21:42.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:42 vm01 ceph-mon[48597]: [10/Mar/2026:10:21:40] ENGINE Serving on https://192.168.123.101:7150 2026-03-10T10:21:42.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:42 vm01 ceph-mon[48597]: [10/Mar/2026:10:21:40] ENGINE Bus STARTED 2026-03-10T10:21:42.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:42 vm01 ceph-mon[48597]: mgrmap e11: vm01.nkkzmc(active, since 1.06256s) 2026-03-10T10:21:42.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:42 vm01 ceph-mon[48597]: from='client.14166 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-10T10:21:42.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:42 vm01 ceph-mon[48597]: from='client.14166 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-10T10:21:42.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:42 vm01 ceph-mon[48597]: from='client.14174 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:21:42.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:42 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:42.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:42 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:42.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:42 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:42.420 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: 8443 2026-03-10T10:21:42.454 INFO:teuthology.orchestra.run.vm01.stderr:firewalld does not appear to be present 2026-03-10T10:21:42.454 INFO:teuthology.orchestra.run.vm01.stderr:Not possible to open ports <[8443]>. firewalld.service is not available 2026-03-10T10:21:42.455 INFO:teuthology.orchestra.run.vm01.stderr:Ceph Dashboard is now available at: 2026-03-10T10:21:42.455 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-10T10:21:42.456 INFO:teuthology.orchestra.run.vm01.stderr: URL: https://vm01.local:8443/ 2026-03-10T10:21:42.456 INFO:teuthology.orchestra.run.vm01.stderr: User: admin 2026-03-10T10:21:42.456 INFO:teuthology.orchestra.run.vm01.stderr: Password: lizekbph8k 2026-03-10T10:21:42.456 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-10T10:21:42.456 INFO:teuthology.orchestra.run.vm01.stderr:Enabling autotune for osd_memory_target 2026-03-10T10:21:43.079 INFO:teuthology.orchestra.run.vm01.stderr:/usr/bin/ceph: set mgr/dashboard/cluster/status 2026-03-10T10:21:43.108 INFO:teuthology.orchestra.run.vm01.stderr:You can access the Ceph CLI with: 2026-03-10T10:21:43.108 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-10T10:21:43.108 INFO:teuthology.orchestra.run.vm01.stderr: sudo /home/ubuntu/cephtest/cephadm shell --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring 2026-03-10T10:21:43.108 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-10T10:21:43.108 INFO:teuthology.orchestra.run.vm01.stderr:Please consider enabling telemetry to help improve Ceph: 2026-03-10T10:21:43.108 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-10T10:21:43.108 INFO:teuthology.orchestra.run.vm01.stderr: ceph telemetry on 2026-03-10T10:21:43.108 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-10T10:21:43.108 INFO:teuthology.orchestra.run.vm01.stderr:For more information see: 2026-03-10T10:21:43.108 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-10T10:21:43.109 INFO:teuthology.orchestra.run.vm01.stderr: https://docs.ceph.com/docs/master/mgr/telemetry/ 2026-03-10T10:21:43.109 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-10T10:21:43.109 INFO:teuthology.orchestra.run.vm01.stderr:Bootstrap complete. 2026-03-10T10:21:43.149 INFO:tasks.cephadm:Fetching config... 2026-03-10T10:21:43.149 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T10:21:43.149 DEBUG:teuthology.orchestra.run.vm01:> dd if=/etc/ceph/ceph.conf of=/dev/stdout 2026-03-10T10:21:43.179 INFO:tasks.cephadm:Fetching client.admin keyring... 2026-03-10T10:21:43.179 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T10:21:43.179 DEBUG:teuthology.orchestra.run.vm01:> dd if=/etc/ceph/ceph.client.admin.keyring of=/dev/stdout 2026-03-10T10:21:43.238 INFO:tasks.cephadm:Fetching mon keyring... 2026-03-10T10:21:43.238 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T10:21:43.238 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/var/lib/ceph/a6b3e19a-1c6a-11f1-aaba-ff800f171c2f/mon.vm01/keyring of=/dev/stdout 2026-03-10T10:21:43.308 INFO:tasks.cephadm:Fetching pub ssh key... 2026-03-10T10:21:43.308 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T10:21:43.308 DEBUG:teuthology.orchestra.run.vm01:> dd if=/home/ubuntu/cephtest/ceph.pub of=/dev/stdout 2026-03-10T10:21:43.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:43 vm01 ceph-mon[48597]: from='client.14176 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:21:43.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:43 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/1187779892' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-10T10:21:43.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:43 vm01 ceph-mon[48597]: mgrmap e12: vm01.nkkzmc(active, since 2s) 2026-03-10T10:21:43.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:43 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/2913653495' entity='client.admin' 2026-03-10T10:21:43.372 INFO:tasks.cephadm:Installing pub ssh key for root users... 2026-03-10T10:21:43.372 DEBUG:teuthology.orchestra.run.vm01:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDRXCUjFz3/Mho9h06Q/GCm4LjUEOOOe1LThXg5SOOwvJ6pBA0rhao9Bpq83CsmV47txnlcFQaalL9kjvV7HhDMtU5+nKti1O61CoSaYbOU4DBMGDrkEzpF+56LCQjJHL9CCtEvPjlIDuogGrN3mg8Xplss7Wb0+ckT5hqMST+vguEFhUrkpr3a3GVAGqGY/tgUDrufgNc96P3GPKpUDZesAA+AMvbhoFF++mnMly9NR8TlLHEKT0k96FJDM4XdudWc2R82//PTJPhCT+npxn7qUaBTDkmYL/R+RlpNuNfsdyxDcJso6EzKoM4wKJASr8UBS6LhCt680X6C41Gm/DLhZqG1Rh8oaQrq7nAb0zh3OxlVYGCgvuBl1RVMyq9U5sQJMiLS1AxrFSKvociO6oNsLFunV3VCQUKdm0UeGrs0aSP8l/2s0a/qxxtKBvrjgtPNqtONPBlpxbCJIyT6tLivbLvZT32NFYRuJN8kvFYB5gyQYRjMENVXO4NHoBeoa5c= ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-10T10:21:43.456 INFO:teuthology.orchestra.run.vm01.stdout:ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDRXCUjFz3/Mho9h06Q/GCm4LjUEOOOe1LThXg5SOOwvJ6pBA0rhao9Bpq83CsmV47txnlcFQaalL9kjvV7HhDMtU5+nKti1O61CoSaYbOU4DBMGDrkEzpF+56LCQjJHL9CCtEvPjlIDuogGrN3mg8Xplss7Wb0+ckT5hqMST+vguEFhUrkpr3a3GVAGqGY/tgUDrufgNc96P3GPKpUDZesAA+AMvbhoFF++mnMly9NR8TlLHEKT0k96FJDM4XdudWc2R82//PTJPhCT+npxn7qUaBTDkmYL/R+RlpNuNfsdyxDcJso6EzKoM4wKJASr8UBS6LhCt680X6C41Gm/DLhZqG1Rh8oaQrq7nAb0zh3OxlVYGCgvuBl1RVMyq9U5sQJMiLS1AxrFSKvociO6oNsLFunV3VCQUKdm0UeGrs0aSP8l/2s0a/qxxtKBvrjgtPNqtONPBlpxbCJIyT6tLivbLvZT32NFYRuJN8kvFYB5gyQYRjMENVXO4NHoBeoa5c= ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f 2026-03-10T10:21:43.469 DEBUG:teuthology.orchestra.run.vm08:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDRXCUjFz3/Mho9h06Q/GCm4LjUEOOOe1LThXg5SOOwvJ6pBA0rhao9Bpq83CsmV47txnlcFQaalL9kjvV7HhDMtU5+nKti1O61CoSaYbOU4DBMGDrkEzpF+56LCQjJHL9CCtEvPjlIDuogGrN3mg8Xplss7Wb0+ckT5hqMST+vguEFhUrkpr3a3GVAGqGY/tgUDrufgNc96P3GPKpUDZesAA+AMvbhoFF++mnMly9NR8TlLHEKT0k96FJDM4XdudWc2R82//PTJPhCT+npxn7qUaBTDkmYL/R+RlpNuNfsdyxDcJso6EzKoM4wKJASr8UBS6LhCt680X6C41Gm/DLhZqG1Rh8oaQrq7nAb0zh3OxlVYGCgvuBl1RVMyq9U5sQJMiLS1AxrFSKvociO6oNsLFunV3VCQUKdm0UeGrs0aSP8l/2s0a/qxxtKBvrjgtPNqtONPBlpxbCJIyT6tLivbLvZT32NFYRuJN8kvFYB5gyQYRjMENVXO4NHoBeoa5c= ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-10T10:21:43.510 INFO:teuthology.orchestra.run.vm08.stdout:ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDRXCUjFz3/Mho9h06Q/GCm4LjUEOOOe1LThXg5SOOwvJ6pBA0rhao9Bpq83CsmV47txnlcFQaalL9kjvV7HhDMtU5+nKti1O61CoSaYbOU4DBMGDrkEzpF+56LCQjJHL9CCtEvPjlIDuogGrN3mg8Xplss7Wb0+ckT5hqMST+vguEFhUrkpr3a3GVAGqGY/tgUDrufgNc96P3GPKpUDZesAA+AMvbhoFF++mnMly9NR8TlLHEKT0k96FJDM4XdudWc2R82//PTJPhCT+npxn7qUaBTDkmYL/R+RlpNuNfsdyxDcJso6EzKoM4wKJASr8UBS6LhCt680X6C41Gm/DLhZqG1Rh8oaQrq7nAb0zh3OxlVYGCgvuBl1RVMyq9U5sQJMiLS1AxrFSKvociO6oNsLFunV3VCQUKdm0UeGrs0aSP8l/2s0a/qxxtKBvrjgtPNqtONPBlpxbCJIyT6tLivbLvZT32NFYRuJN8kvFYB5gyQYRjMENVXO4NHoBeoa5c= ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f 2026-03-10T10:21:43.521 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph config set mgr mgr/cephadm/allow_ptrace true 2026-03-10T10:21:44.229 INFO:tasks.cephadm:Distributing conf and client.admin keyring to all hosts + 0755 2026-03-10T10:21:44.230 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph orch client-keyring set client.admin '*' --mode 0755 2026-03-10T10:21:44.808 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm08 2026-03-10T10:21:44.808 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-10T10:21:44.808 DEBUG:teuthology.orchestra.run.vm08:> dd of=/etc/ceph/ceph.conf 2026-03-10T10:21:44.824 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-10T10:21:44.824 DEBUG:teuthology.orchestra.run.vm08:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:21:44.878 INFO:tasks.cephadm:Adding host vm08 to orchestrator... 2026-03-10T10:21:44.878 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph orch host add vm08 2026-03-10T10:21:45.353 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:45 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/586187444' entity='client.admin' 2026-03-10T10:21:45.353 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:45 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:45.353 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:45 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:45.353 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:45 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:21:45.353 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:45 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:46.099 INFO:teuthology.orchestra.run.vm01.stdout:Added host 'vm08' with addr '192.168.123.108' 2026-03-10T10:21:46.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:46 vm01 ceph-mon[48597]: from='client.14186 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:21:46.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:46 vm01 ceph-mon[48597]: Deploying daemon alertmanager.vm01 on vm01 2026-03-10T10:21:46.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:46 vm01 ceph-mon[48597]: from='client.14188 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm08", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:21:46.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:46 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:46.781 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph orch host ls --format=json 2026-03-10T10:21:47.387 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:47 vm01 ceph-mon[48597]: Deploying cephadm binary to vm08 2026-03-10T10:21:47.387 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:47 vm01 ceph-mon[48597]: Added host vm08 2026-03-10T10:21:47.387 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:47 vm01 ceph-mon[48597]: mgrmap e13: vm01.nkkzmc(active, since 6s) 2026-03-10T10:21:47.387 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T10:21:47.387 INFO:teuthology.orchestra.run.vm01.stdout:[{"addr": "192.168.123.101", "hostname": "vm01", "labels": [], "status": ""}, {"addr": "192.168.123.108", "hostname": "vm08", "labels": [], "status": ""}] 2026-03-10T10:21:47.455 INFO:tasks.cephadm:Setting crush tunables to default 2026-03-10T10:21:47.455 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph osd crush tunables default 2026-03-10T10:21:48.292 INFO:teuthology.orchestra.run.vm01.stderr:adjusted tunables profile to default 2026-03-10T10:21:48.500 INFO:tasks.cephadm:Adding mon.vm01 on vm01 2026-03-10T10:21:48.500 INFO:tasks.cephadm:Adding mon.vm08 on vm08 2026-03-10T10:21:48.500 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph orch apply mon '2;vm01:192.168.123.101=vm01;vm08:192.168.123.108=vm08' 2026-03-10T10:21:48.582 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:48 vm01 ceph-mon[48597]: from='client.14190 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:21:48.582 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:48 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/1672009625' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-10T10:21:49.001 INFO:teuthology.orchestra.run.vm08.stdout:Scheduled mon update... 2026-03-10T10:21:49.056 DEBUG:teuthology.orchestra.run.vm08:mon.vm08> sudo journalctl -f -n 0 -u ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f@mon.vm08.service 2026-03-10T10:21:49.058 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:21:49.058 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph mon dump -f json 2026-03-10T10:21:49.430 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:49 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/1672009625' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-10T10:21:49.430 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:49 vm01 ceph-mon[48597]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T10:21:49.430 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:49 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:49.430 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:49 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm01", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-10T10:21:49.430 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:49 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "auth get-or-create", "entity": "client.crash.vm01", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]': finished 2026-03-10T10:21:49.430 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:49 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:21:49.430 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:49 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:49.611 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-10T10:21:49.611 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":1,"fsid":"a6b3e19a-1c6a-11f1-aaba-ff800f171c2f","modified":"2026-03-10T10:21:11.462968Z","created":"2026-03-10T10:21:11.462968Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:21:49.613 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 1 2026-03-10T10:21:50.679 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:21:50.679 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph mon dump -f json 2026-03-10T10:21:50.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:50 vm01 ceph-mon[48597]: Deploying daemon crash.vm01 on vm01 2026-03-10T10:21:50.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:50 vm01 ceph-mon[48597]: from='client.14194 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "2;vm01:192.168.123.101=vm01;vm08:192.168.123.108=vm08", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:21:50.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:50 vm01 ceph-mon[48597]: Saving service mon spec with placement vm01:192.168.123.101=vm01;vm08:192.168.123.108=vm08;count:2 2026-03-10T10:21:50.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:50 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:50.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:50 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:50.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:50 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:50.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:50 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "dashboard set-grafana-api-ssl-verify", "value": "false"}]: dispatch 2026-03-10T10:21:50.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:50 vm01 ceph-mon[48597]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-ssl-verify", "value": "false"}]: dispatch 2026-03-10T10:21:50.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:50 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:50.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:50 vm01 ceph-mon[48597]: Deploying daemon grafana.vm01 on vm01 2026-03-10T10:21:50.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:50 vm01 ceph-mon[48597]: from='client.? 192.168.123.108:0/3684017951' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:50.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:50 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' 2026-03-10T10:21:51.464 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-10T10:21:51.464 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":1,"fsid":"a6b3e19a-1c6a-11f1-aaba-ff800f171c2f","modified":"2026-03-10T10:21:11.462968Z","created":"2026-03-10T10:21:11.462968Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:21:51.466 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 1 2026-03-10T10:21:51.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:51 vm01 ceph-mon[48597]: from='client.? 192.168.123.108:0/1780431496' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:52.518 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:21:52.518 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph mon dump -f json 2026-03-10T10:21:53.032 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-10T10:21:53.032 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":1,"fsid":"a6b3e19a-1c6a-11f1-aaba-ff800f171c2f","modified":"2026-03-10T10:21:11.462968Z","created":"2026-03-10T10:21:11.462968Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:21:53.034 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 1 2026-03-10T10:21:53.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:53 vm01 ceph-mon[48597]: from='client.? 192.168.123.108:0/1814267629' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:54.096 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:21:54.096 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph mon dump -f json 2026-03-10T10:21:54.613 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-10T10:21:54.613 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":1,"fsid":"a6b3e19a-1c6a-11f1-aaba-ff800f171c2f","modified":"2026-03-10T10:21:11.462968Z","created":"2026-03-10T10:21:11.462968Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:21:54.616 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 1 2026-03-10T10:21:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:54 vm01 ceph-mon[48597]: from='client.? 192.168.123.108:0/2990321405' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:55.783 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:21:55.783 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph mon dump -f json 2026-03-10T10:21:56.353 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-10T10:21:56.354 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":1,"fsid":"a6b3e19a-1c6a-11f1-aaba-ff800f171c2f","modified":"2026-03-10T10:21:11.462968Z","created":"2026-03-10T10:21:11.462968Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:21:56.356 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 1 2026-03-10T10:21:56.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:56 vm01 ceph-mon[48597]: from='client.? 192.168.123.108:0/4230829633' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:57.430 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:21:57.430 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph mon dump -f json 2026-03-10T10:21:58.111 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-10T10:21:58.111 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":1,"fsid":"a6b3e19a-1c6a-11f1-aaba-ff800f171c2f","modified":"2026-03-10T10:21:11.462968Z","created":"2026-03-10T10:21:11.462968Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:21:58.115 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 1 2026-03-10T10:21:58.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:21:58 vm01 ceph-mon[48597]: from='client.? 192.168.123.108:0/1469809813' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:59.182 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:21:59.182 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph mon dump -f json 2026-03-10T10:21:59.702 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-10T10:21:59.702 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":1,"fsid":"a6b3e19a-1c6a-11f1-aaba-ff800f171c2f","modified":"2026-03-10T10:21:11.462968Z","created":"2026-03-10T10:21:11.462968Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:21:59.705 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 1 2026-03-10T10:22:00.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:00 vm01 ceph-mon[48597]: from='client.? 192.168.123.108:0/600667728' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:22:00.770 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:22:00.770 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph mon dump -f json 2026-03-10T10:22:01.422 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-10T10:22:01.422 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":1,"fsid":"a6b3e19a-1c6a-11f1-aaba-ff800f171c2f","modified":"2026-03-10T10:21:11.462968Z","created":"2026-03-10T10:21:11.462968Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:22:01.425 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 1 2026-03-10T10:22:01.461 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:01 vm01 ceph-mon[48597]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:22:02.472 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:22:02.472 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph mon dump -f json 2026-03-10T10:22:02.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:02 vm01 ceph-mon[48597]: from='client.? 192.168.123.108:0/467850947' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:22:02.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:02 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:03.020 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-10T10:22:03.020 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":1,"fsid":"a6b3e19a-1c6a-11f1-aaba-ff800f171c2f","modified":"2026-03-10T10:21:11.462968Z","created":"2026-03-10T10:21:11.462968Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:22:03.022 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 1 2026-03-10T10:22:03.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:03 vm01 ceph-mon[48597]: Deploying daemon node-exporter.vm01 on vm01 2026-03-10T10:22:03.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:03 vm01 ceph-mon[48597]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:22:03.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:03 vm01 ceph-mon[48597]: from='client.? 192.168.123.108:0/2079813961' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:22:04.111 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:22:04.111 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph mon dump -f json 2026-03-10T10:22:04.623 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-10T10:22:04.623 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":1,"fsid":"a6b3e19a-1c6a-11f1-aaba-ff800f171c2f","modified":"2026-03-10T10:21:11.462968Z","created":"2026-03-10T10:21:11.462968Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:22:04.624 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 1 2026-03-10T10:22:05.687 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:22:05.687 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph mon dump -f json 2026-03-10T10:22:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:05 vm01 ceph-mon[48597]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:22:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:05 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:05 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mgr module enable", "module": "prometheus"}]: dispatch 2026-03-10T10:22:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:05 vm01 ceph-mon[48597]: from='client.? 192.168.123.108:0/2075133132' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:22:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:05 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:06.206 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-10T10:22:06.206 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":1,"fsid":"a6b3e19a-1c6a-11f1-aaba-ff800f171c2f","modified":"2026-03-10T10:21:11.462968Z","created":"2026-03-10T10:21:11.462968Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:22:06.208 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 1 2026-03-10T10:22:07.263 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:06 vm01 ceph-mon[48597]: from='mgr.14162 192.168.123.101:0/3402891340' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "mgr module enable", "module": "prometheus"}]': finished 2026-03-10T10:22:07.263 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:06 vm01 ceph-mon[48597]: mgrmap e14: vm01.nkkzmc(active, since 25s) 2026-03-10T10:22:07.263 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:06 vm01 ceph-mon[48597]: from='client.? 192.168.123.108:0/2118626765' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:22:07.951 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:22:07.951 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph mon dump -f json 2026-03-10T10:22:08.481 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-10T10:22:08.481 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":1,"fsid":"a6b3e19a-1c6a-11f1-aaba-ff800f171c2f","modified":"2026-03-10T10:21:11.462968Z","created":"2026-03-10T10:21:11.462968Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:22:08.483 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 1 2026-03-10T10:22:08.822 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:08 vm01 ceph-mon[48597]: from='client.? 192.168.123.108:0/2210260630' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:22:09.548 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:22:09.549 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph mon dump -f json 2026-03-10T10:22:10.047 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-10T10:22:10.047 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":1,"fsid":"a6b3e19a-1c6a-11f1-aaba-ff800f171c2f","modified":"2026-03-10T10:21:11.462968Z","created":"2026-03-10T10:21:11.462968Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:22:10.050 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 1 2026-03-10T10:22:10.818 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:10 vm01 ceph-mon[48597]: Active manager daemon vm01.nkkzmc restarted 2026-03-10T10:22:10.819 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:10 vm01 ceph-mon[48597]: Activating manager daemon vm01.nkkzmc 2026-03-10T10:22:10.819 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:10 vm01 ceph-mon[48597]: osdmap e5: 0 total, 0 up, 0 in 2026-03-10T10:22:10.819 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:10 vm01 ceph-mon[48597]: mgrmap e15: vm01.nkkzmc(active, starting, since 0.421321s) 2026-03-10T10:22:10.819 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-10T10:22:10.819 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mgr metadata", "who": "vm01.nkkzmc", "id": "vm01.nkkzmc"}]: dispatch 2026-03-10T10:22:10.819 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T10:22:10.819 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T10:22:10.819 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T10:22:10.819 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:10 vm01 ceph-mon[48597]: Manager daemon vm01.nkkzmc is now available 2026-03-10T10:22:10.819 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:10 vm01 ceph-mon[48597]: from='client.? 192.168.123.108:0/3854175452' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:22:10.819 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:10.819 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:22:10.819 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:22:10.819 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:10.819 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:22:10.819 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:22:10.819 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:22:10.819 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:11.100 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:22:11.100 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph mon dump -f json 2026-03-10T10:22:11.718 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-10T10:22:11.718 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":1,"fsid":"a6b3e19a-1c6a-11f1-aaba-ff800f171c2f","modified":"2026-03-10T10:21:11.462968Z","created":"2026-03-10T10:21:11.462968Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:22:11.723 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 1 2026-03-10T10:22:11.869 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:11 vm01 ceph-mon[48597]: [10/Mar/2026:10:22:10] ENGINE Bus STARTING 2026-03-10T10:22:11.869 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:11 vm01 ceph-mon[48597]: [10/Mar/2026:10:22:10] ENGINE Serving on https://192.168.123.101:7150 2026-03-10T10:22:11.869 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:11 vm01 ceph-mon[48597]: [10/Mar/2026:10:22:10] ENGINE Bus STARTED 2026-03-10T10:22:11.869 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:11.869 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:11.869 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:11 vm01 ceph-mon[48597]: mgrmap e16: vm01.nkkzmc(active, since 1.42682s) 2026-03-10T10:22:11.869 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:11.869 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:11.869 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:11.869 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:22:12.813 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:22:12.813 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph mon dump -f json 2026-03-10T10:22:13.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:12 vm01 ceph-mon[48597]: Updating vm01:/etc/ceph/ceph.conf 2026-03-10T10:22:13.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:12 vm01 ceph-mon[48597]: from='client.? 192.168.123.108:0/783780241' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:22:13.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:12 vm01 ceph-mon[48597]: Updating vm01:/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:22:13.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:12 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:13.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:12 vm01 ceph-mon[48597]: mgrmap e17: vm01.nkkzmc(active, since 2s) 2026-03-10T10:22:13.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:12 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:13.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:12 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:13.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:12 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:22:13.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:12 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:13.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:12 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm08", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-10T10:22:13.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:12 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "auth get-or-create", "entity": "client.crash.vm08", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]': finished 2026-03-10T10:22:13.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:12 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:13.868 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-10T10:22:13.868 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":1,"fsid":"a6b3e19a-1c6a-11f1-aaba-ff800f171c2f","modified":"2026-03-10T10:21:11.462968Z","created":"2026-03-10T10:21:11.462968Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:22:13.873 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 1 2026-03-10T10:22:14.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:13 vm01 ceph-mon[48597]: Deploying daemon crash.vm08 on vm08 2026-03-10T10:22:14.953 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:22:14.954 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph mon dump -f json 2026-03-10T10:22:15.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:14 vm01 ceph-mon[48597]: from='client.? 192.168.123.108:0/4066588419' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:22:15.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:14 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:15.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:14 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm08.rwpjaf", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:22:15.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:14 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.vm08.rwpjaf", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-10T10:22:15.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:14 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:22:15.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:14 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:15.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:14 vm01 ceph-mon[48597]: Deploying daemon mgr.vm08.rwpjaf on vm08 2026-03-10T10:22:15.907 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-10T10:22:15.907 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":1,"fsid":"a6b3e19a-1c6a-11f1-aaba-ff800f171c2f","modified":"2026-03-10T10:21:11.462968Z","created":"2026-03-10T10:21:11.462968Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:22:15.914 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 1 2026-03-10T10:22:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:15 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:15 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:22:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:15 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:15 vm01 ceph-mon[48597]: Deploying daemon mon.vm08 on vm08 2026-03-10T10:22:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:15 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:17.036 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:22:17.036 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph mon dump -f json 2026-03-10T10:22:21.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:21 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:21 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-10T10:22:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:21 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-10T10:22:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:21 vm01 ceph-mon[48597]: mon.vm01 calling monitor election 2026-03-10T10:22:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:21 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-10T10:22:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:21 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-10T10:22:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:21 vm01 ceph-mon[48597]: mon.vm08 calling monitor election 2026-03-10T10:22:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:21 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-10T10:22:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:21 vm01 ceph-mon[48597]: from='mgr.? 192.168.123.108:0/121138284' entity='mgr.vm08.rwpjaf' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm08.rwpjaf/crt"}]: dispatch 2026-03-10T10:22:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:21 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-10T10:22:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:21 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-10T10:22:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:21 vm01 ceph-mon[48597]: mon.vm01 is new leader, mons vm01,vm08 in quorum (ranks 0,1) 2026-03-10T10:22:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:21 vm01 ceph-mon[48597]: monmap e2: 2 mons at {vm01=[v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0],vm08=[v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0]} 2026-03-10T10:22:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:21 vm01 ceph-mon[48597]: fsmap 2026-03-10T10:22:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:21 vm01 ceph-mon[48597]: osdmap e5: 0 total, 0 up, 0 in 2026-03-10T10:22:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:21 vm01 ceph-mon[48597]: mgrmap e17: vm01.nkkzmc(active, since 11s) 2026-03-10T10:22:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:21 vm01 ceph-mon[48597]: Standby manager daemon vm08.rwpjaf started 2026-03-10T10:22:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:21 vm01 ceph-mon[48597]: from='mgr.? 192.168.123.108:0/121138284' entity='mgr.vm08.rwpjaf' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-10T10:22:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:21 vm01 ceph-mon[48597]: overall HEALTH_OK 2026-03-10T10:22:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:21 vm01 ceph-mon[48597]: from='mgr.? 192.168.123.108:0/121138284' entity='mgr.vm08.rwpjaf' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm08.rwpjaf/key"}]: dispatch 2026-03-10T10:22:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:21 vm01 ceph-mon[48597]: from='mgr.? 192.168.123.108:0/121138284' entity='mgr.vm08.rwpjaf' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-10T10:22:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:21 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:21 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:22.014 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-10T10:22:22.014 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":2,"fsid":"a6b3e19a-1c6a-11f1-aaba-ff800f171c2f","modified":"2026-03-10T10:22:16.457288Z","created":"2026-03-10T10:21:11.462968Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"vm08","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1]} 2026-03-10T10:22:22.016 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 2 2026-03-10T10:22:22.078 INFO:tasks.cephadm:Generating final ceph.conf file... 2026-03-10T10:22:22.078 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph config generate-minimal-conf 2026-03-10T10:22:22.641 INFO:teuthology.orchestra.run.vm01.stdout:# minimal ceph.conf for a6b3e19a-1c6a-11f1-aaba-ff800f171c2f 2026-03-10T10:22:22.641 INFO:teuthology.orchestra.run.vm01.stdout:[global] 2026-03-10T10:22:22.641 INFO:teuthology.orchestra.run.vm01.stdout: fsid = a6b3e19a-1c6a-11f1-aaba-ff800f171c2f 2026-03-10T10:22:22.641 INFO:teuthology.orchestra.run.vm01.stdout: mon_host = [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] 2026-03-10T10:22:22.653 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:22 vm01 ceph-mon[48597]: Deploying daemon prometheus.vm01 on vm01 2026-03-10T10:22:22.653 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:22 vm01 ceph-mon[48597]: mgrmap e18: vm01.nkkzmc(active, since 11s), standbys: vm08.rwpjaf 2026-03-10T10:22:22.653 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:22 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mgr metadata", "who": "vm08.rwpjaf", "id": "vm08.rwpjaf"}]: dispatch 2026-03-10T10:22:22.653 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:22 vm01 ceph-mon[48597]: from='client.? 192.168.123.108:0/1208972660' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:22:22.653 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:22 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-10T10:22:22.702 INFO:tasks.cephadm:Distributing (final) config and client.admin keyring... 2026-03-10T10:22:22.702 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T10:22:22.702 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/ceph/ceph.conf 2026-03-10T10:22:22.726 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T10:22:22.727 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:22:22.792 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-10T10:22:22.793 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/ceph/ceph.conf 2026-03-10T10:22:22.818 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:22 vm08 ceph-mon[50506]: Deploying daemon prometheus.vm01 on vm01 2026-03-10T10:22:22.819 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:22 vm08 ceph-mon[50506]: mgrmap e18: vm01.nkkzmc(active, since 11s), standbys: vm08.rwpjaf 2026-03-10T10:22:22.819 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:22 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mgr metadata", "who": "vm08.rwpjaf", "id": "vm08.rwpjaf"}]: dispatch 2026-03-10T10:22:22.819 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:22 vm08 ceph-mon[50506]: from='client.? 192.168.123.108:0/1208972660' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:22:22.819 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:22 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-10T10:22:22.823 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-10T10:22:22.823 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:22:22.889 INFO:tasks.cephadm:Deploying OSDs... 2026-03-10T10:22:22.889 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T10:22:22.889 DEBUG:teuthology.orchestra.run.vm01:> dd if=/scratch_devs of=/dev/stdout 2026-03-10T10:22:22.904 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T10:22:22.905 DEBUG:teuthology.orchestra.run.vm01:> ls /dev/[sv]d? 2026-03-10T10:22:22.960 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vda 2026-03-10T10:22:22.960 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vdb 2026-03-10T10:22:22.960 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vdc 2026-03-10T10:22:22.960 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vdd 2026-03-10T10:22:22.960 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vde 2026-03-10T10:22:22.960 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-10T10:22:22.961 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-10T10:22:22.961 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vdb 2026-03-10T10:22:23.020 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vdb 2026-03-10T10:22:23.020 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T10:22:23.020 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 236 Links: 1 Device type: fc,10 2026-03-10T10:22:23.020 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T10:22:23.020 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T10:22:23.020 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-10 10:21:44.608195429 +0000 2026-03-10T10:22:23.020 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-10 10:21:44.296192978 +0000 2026-03-10T10:22:23.020 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-10 10:21:44.296192978 +0000 2026-03-10T10:22:23.020 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-10 10:16:39.250000000 +0000 2026-03-10T10:22:23.021 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-10T10:22:23.085 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-10T10:22:23.085 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-10T10:22:23.085 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.000158638 s, 3.2 MB/s 2026-03-10T10:22:23.086 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-10T10:22:23.145 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vdc 2026-03-10T10:22:23.204 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vdc 2026-03-10T10:22:23.204 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T10:22:23.204 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 255 Links: 1 Device type: fc,20 2026-03-10T10:22:23.204 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T10:22:23.204 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T10:22:23.204 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-10 10:21:44.665195877 +0000 2026-03-10T10:22:23.204 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-10 10:21:44.301193018 +0000 2026-03-10T10:22:23.204 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-10 10:21:44.301193018 +0000 2026-03-10T10:22:23.204 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-10 10:16:39.257000000 +0000 2026-03-10T10:22:23.205 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-10T10:22:23.268 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-10T10:22:23.269 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-10T10:22:23.269 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.000148038 s, 3.5 MB/s 2026-03-10T10:22:23.269 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-10T10:22:23.327 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vdd 2026-03-10T10:22:23.389 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vdd 2026-03-10T10:22:23.389 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T10:22:23.389 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-10T10:22:23.389 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T10:22:23.389 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T10:22:23.389 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-10 10:21:44.731196395 +0000 2026-03-10T10:22:23.389 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-10 10:21:44.303193033 +0000 2026-03-10T10:22:23.389 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-10 10:21:44.303193033 +0000 2026-03-10T10:22:23.389 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-10 10:16:39.269000000 +0000 2026-03-10T10:22:23.389 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-10T10:22:23.459 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-10T10:22:23.459 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-10T10:22:23.459 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.000228979 s, 2.2 MB/s 2026-03-10T10:22:23.460 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-10T10:22:23.526 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vde 2026-03-10T10:22:23.540 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:23 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/100573125' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:23.559 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vde 2026-03-10T10:22:23.559 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T10:22:23.559 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-10T10:22:23.559 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T10:22:23.559 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T10:22:23.559 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-10 10:21:44.808197000 +0000 2026-03-10T10:22:23.559 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-10 10:21:44.300193010 +0000 2026-03-10T10:22:23.559 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-10 10:21:44.300193010 +0000 2026-03-10T10:22:23.559 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-10 10:16:39.313000000 +0000 2026-03-10T10:22:23.560 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-10T10:22:23.625 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-10T10:22:23.625 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-10T10:22:23.625 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.000122088 s, 4.2 MB/s 2026-03-10T10:22:23.626 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-10T10:22:23.696 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-10T10:22:23.696 DEBUG:teuthology.orchestra.run.vm08:> dd if=/scratch_devs of=/dev/stdout 2026-03-10T10:22:23.713 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T10:22:23.713 DEBUG:teuthology.orchestra.run.vm08:> ls /dev/[sv]d? 2026-03-10T10:22:23.770 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vda 2026-03-10T10:22:23.770 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vdb 2026-03-10T10:22:23.770 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vdc 2026-03-10T10:22:23.770 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vdd 2026-03-10T10:22:23.770 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vde 2026-03-10T10:22:23.770 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-10T10:22:23.770 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-10T10:22:23.771 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vdb 2026-03-10T10:22:23.830 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vdb 2026-03-10T10:22:23.830 INFO:teuthology.orchestra.run.vm08.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T10:22:23.830 INFO:teuthology.orchestra.run.vm08.stdout:Device: 6h/6d Inode: 223 Links: 1 Device type: fc,10 2026-03-10T10:22:23.830 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T10:22:23.830 INFO:teuthology.orchestra.run.vm08.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T10:22:23.830 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-03-10 10:22:12.321586793 +0000 2026-03-10T10:22:23.830 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-03-10 10:22:11.992585903 +0000 2026-03-10T10:22:23.830 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-03-10 10:22:11.992585903 +0000 2026-03-10T10:22:23.830 INFO:teuthology.orchestra.run.vm08.stdout: Birth: 2026-03-10 10:17:05.276000000 +0000 2026-03-10T10:22:23.830 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-10T10:22:23.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:23 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/100573125' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:23.871 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-03-10T10:22:23.871 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-03-10T10:22:23.871 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.000265197 s, 1.9 MB/s 2026-03-10T10:22:23.872 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-10T10:22:23.932 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vdc 2026-03-10T10:22:23.993 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vdc 2026-03-10T10:22:23.993 INFO:teuthology.orchestra.run.vm08.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T10:22:23.993 INFO:teuthology.orchestra.run.vm08.stdout:Device: 6h/6d Inode: 255 Links: 1 Device type: fc,20 2026-03-10T10:22:23.993 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T10:22:23.993 INFO:teuthology.orchestra.run.vm08.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T10:22:23.993 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-03-10 10:22:12.395586993 +0000 2026-03-10T10:22:23.993 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-03-10 10:22:12.001585927 +0000 2026-03-10T10:22:23.993 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-03-10 10:22:12.001585927 +0000 2026-03-10T10:22:23.994 INFO:teuthology.orchestra.run.vm08.stdout: Birth: 2026-03-10 10:17:05.299000000 +0000 2026-03-10T10:22:23.994 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-10T10:22:24.060 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-03-10T10:22:24.060 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-03-10T10:22:24.060 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.000284493 s, 1.8 MB/s 2026-03-10T10:22:24.061 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-10T10:22:24.118 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vdd 2026-03-10T10:22:24.177 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vdd 2026-03-10T10:22:24.177 INFO:teuthology.orchestra.run.vm08.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T10:22:24.177 INFO:teuthology.orchestra.run.vm08.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-10T10:22:24.177 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T10:22:24.177 INFO:teuthology.orchestra.run.vm08.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T10:22:24.177 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-03-10 10:22:12.465587183 +0000 2026-03-10T10:22:24.177 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-03-10 10:22:11.996585914 +0000 2026-03-10T10:22:24.177 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-03-10 10:22:11.996585914 +0000 2026-03-10T10:22:24.177 INFO:teuthology.orchestra.run.vm08.stdout: Birth: 2026-03-10 10:17:05.303000000 +0000 2026-03-10T10:22:24.178 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-10T10:22:24.246 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-03-10T10:22:24.246 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-03-10T10:22:24.246 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.000227385 s, 2.3 MB/s 2026-03-10T10:22:24.247 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-10T10:22:24.304 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vde 2026-03-10T10:22:24.362 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vde 2026-03-10T10:22:24.362 INFO:teuthology.orchestra.run.vm08.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T10:22:24.362 INFO:teuthology.orchestra.run.vm08.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-10T10:22:24.362 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T10:22:24.362 INFO:teuthology.orchestra.run.vm08.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T10:22:24.362 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-03-10 10:22:12.538587381 +0000 2026-03-10T10:22:24.362 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-03-10 10:22:11.997585916 +0000 2026-03-10T10:22:24.362 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-03-10 10:22:11.997585916 +0000 2026-03-10T10:22:24.362 INFO:teuthology.orchestra.run.vm08.stdout: Birth: 2026-03-10 10:17:05.311000000 +0000 2026-03-10T10:22:24.362 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-10T10:22:24.431 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-03-10T10:22:24.431 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-03-10T10:22:24.431 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.0004436 s, 1.2 MB/s 2026-03-10T10:22:24.432 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-10T10:22:24.492 INFO:tasks.cephadm:Deploying osd.0 on vm01 with /dev/vde... 2026-03-10T10:22:24.492 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- lvm zap /dev/vde 2026-03-10T10:22:27.671 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T10:22:27.690 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph orch daemon add osd vm01:/dev/vde 2026-03-10T10:22:28.236 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:28 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:28.236 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:28 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:22:28.236 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:28 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:28.236 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:28 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:22:28.236 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:28 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:28.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:28 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:28.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:28 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:22:28.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:28 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:28.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:28 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:22:28.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:28 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:29 vm08 ceph-mon[50506]: Updating vm08:/etc/ceph/ceph.conf 2026-03-10T10:22:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:29 vm08 ceph-mon[50506]: Updating vm08:/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:22:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:29 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:29 vm08 ceph-mon[50506]: from='client.24101 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:22:29.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:29 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:22:29.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:29 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:22:29.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:29 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:29.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:29 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:29.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:29 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:29.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:29 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:29.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:29 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:22:29.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:29 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:22:29.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:29 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:29.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:29 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/2798059938' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "ebe7975c-d06c-485f-8d2b-ebf217684065"}]: dispatch 2026-03-10T10:22:29.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:29 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/2798059938' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "ebe7975c-d06c-485f-8d2b-ebf217684065"}]': finished 2026-03-10T10:22:29.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:29 vm08 ceph-mon[50506]: osdmap e6: 1 total, 0 up, 1 in 2026-03-10T10:22:29.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:29 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:22:29.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:29 vm01 ceph-mon[48597]: Updating vm08:/etc/ceph/ceph.conf 2026-03-10T10:22:29.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:29 vm01 ceph-mon[48597]: Updating vm08:/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:22:29.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:29 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:29.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:29 vm01 ceph-mon[48597]: from='client.24101 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:22:29.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:29 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:22:29.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:29 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:22:29.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:29 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:29.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:29 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:29.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:29 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:29.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:29 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:29.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:29 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:22:29.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:29 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:22:29.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:29 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:29.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:29 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/2798059938' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "ebe7975c-d06c-485f-8d2b-ebf217684065"}]: dispatch 2026-03-10T10:22:29.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:29 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/2798059938' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "ebe7975c-d06c-485f-8d2b-ebf217684065"}]': finished 2026-03-10T10:22:29.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:29 vm01 ceph-mon[48597]: osdmap e6: 1 total, 0 up, 1 in 2026-03-10T10:22:29.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:29 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:22:30.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:30 vm08 ceph-mon[50506]: Updating vm01:/etc/ceph/ceph.conf 2026-03-10T10:22:30.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:30 vm08 ceph-mon[50506]: Reconfiguring mon.vm01 (unknown last config time)... 2026-03-10T10:22:30.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:30 vm08 ceph-mon[50506]: Reconfiguring daemon mon.vm01 on vm01 2026-03-10T10:22:30.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:30 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/1504206345' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:22:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:30 vm01 ceph-mon[48597]: Updating vm01:/etc/ceph/ceph.conf 2026-03-10T10:22:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:30 vm01 ceph-mon[48597]: Reconfiguring mon.vm01 (unknown last config time)... 2026-03-10T10:22:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:30 vm01 ceph-mon[48597]: Reconfiguring daemon mon.vm01 on vm01 2026-03-10T10:22:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:30 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/1504206345' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:22:32.036 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:31 vm01 ceph-mon[48597]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:22:32.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:31 vm08 ceph-mon[50506]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:22:33.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:32 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:33.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:32 vm01 ceph-mon[48597]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:22:33.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:32 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:33.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:32 vm01 ceph-mon[48597]: Reconfiguring mgr.vm01.nkkzmc (unknown last config time)... 2026-03-10T10:22:33.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:32 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm01.nkkzmc", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:22:33.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:32 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:22:33.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:32 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:33.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:32 vm01 ceph-mon[48597]: Reconfiguring daemon mgr.vm01.nkkzmc on vm01 2026-03-10T10:22:33.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:32 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:33.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:32 vm08 ceph-mon[50506]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:22:33.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:32 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:33.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:32 vm08 ceph-mon[50506]: Reconfiguring mgr.vm01.nkkzmc (unknown last config time)... 2026-03-10T10:22:33.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:32 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm01.nkkzmc", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:22:33.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:32 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:22:33.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:32 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:33.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:32 vm08 ceph-mon[50506]: Reconfiguring daemon mgr.vm01.nkkzmc on vm01 2026-03-10T10:22:34.196 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:33 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:34.196 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:33 vm01 ceph-mon[48597]: Reconfiguring alertmanager.vm01 (dependencies changed)... 2026-03-10T10:22:34.196 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:33 vm01 ceph-mon[48597]: Reconfiguring daemon alertmanager.vm01 on vm01 2026-03-10T10:22:34.196 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:33 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T10:22:34.196 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:33 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:34.197 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:33 vm01 ceph-mon[48597]: Deploying daemon osd.0 on vm01 2026-03-10T10:22:34.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:33 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:34.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:33 vm08 ceph-mon[50506]: Reconfiguring alertmanager.vm01 (dependencies changed)... 2026-03-10T10:22:34.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:33 vm08 ceph-mon[50506]: Reconfiguring daemon alertmanager.vm01 on vm01 2026-03-10T10:22:34.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:33 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T10:22:34.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:33 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:34.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:33 vm08 ceph-mon[50506]: Deploying daemon osd.0 on vm01 2026-03-10T10:22:35.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:35 vm01 ceph-mon[48597]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:22:35.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:35 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:35.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:35 vm01 ceph-mon[48597]: Reconfiguring crash.vm01 (monmap changed)... 2026-03-10T10:22:35.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:35 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm01", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-10T10:22:35.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:35 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:35.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:35 vm01 ceph-mon[48597]: Reconfiguring daemon crash.vm01 on vm01 2026-03-10T10:22:35.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:35 vm08 ceph-mon[50506]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:22:35.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:35 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:35.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:35 vm08 ceph-mon[50506]: Reconfiguring crash.vm01 (monmap changed)... 2026-03-10T10:22:35.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:35 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm01", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-10T10:22:35.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:35 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:35.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:35 vm08 ceph-mon[50506]: Reconfiguring daemon crash.vm01 on vm01 2026-03-10T10:22:36.563 INFO:teuthology.orchestra.run.vm01.stdout:Created osd(s) 0 on host 'vm01' 2026-03-10T10:22:36.616 DEBUG:teuthology.orchestra.run.vm01:osd.0> sudo journalctl -f -n 0 -u ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f@osd.0.service 2026-03-10T10:22:36.617 INFO:tasks.cephadm:Deploying osd.1 on vm01 with /dev/vdd... 2026-03-10T10:22:36.617 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- lvm zap /dev/vdd 2026-03-10T10:22:37.041 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:36 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:37.041 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:36 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:37.041 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:36 vm01 ceph-mon[48597]: Reconfiguring grafana.vm01 (dependencies changed)... 2026-03-10T10:22:37.041 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:36 vm01 ceph-mon[48597]: Reconfiguring daemon grafana.vm01 on vm01 2026-03-10T10:22:37.041 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:36 vm01 ceph-mon[48597]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:22:37.041 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:36 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:37.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:36 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:37.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:36 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:37.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:36 vm08 ceph-mon[50506]: Reconfiguring grafana.vm01 (dependencies changed)... 2026-03-10T10:22:37.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:36 vm08 ceph-mon[50506]: Reconfiguring daemon grafana.vm01 on vm01 2026-03-10T10:22:37.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:36 vm08 ceph-mon[50506]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:22:37.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:36 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:37.566 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T10:22:37.582 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph orch daemon add osd vm01:/dev/vdd 2026-03-10T10:22:38.118 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 10:22:37 vm01 ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f-osd-0[58396]: 2026-03-10T10:22:37.879+0000 7fd5d51fa3c0 -1 osd.0 0 log_to_monitors true 2026-03-10T10:22:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: Reconfiguring crash.vm08 (monmap changed)... 2026-03-10T10:22:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm08", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-10T10:22:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: Reconfiguring daemon crash.vm08 on vm08 2026-03-10T10:22:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:38.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm08.rwpjaf", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:22:38.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:22:38.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:38.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:38.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:22:38.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:22:38.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:38.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: from='osd.0 [v2:192.168.123.101:6802/3587270067,v1:192.168.123.101:6803/3587270067]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T10:22:38.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:38.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-10T10:22:38.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://192.168.123.101:9093"}]: dispatch 2026-03-10T10:22:38.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:38.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-10T10:22:38.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://192.168.123.101:3000"}]: dispatch 2026-03-10T10:22:38.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:38.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-10T10:22:38.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://192.168.123.101:9095"}]: dispatch 2026-03-10T10:22:38.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:38.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:22:38.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:38.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:22:38.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:22:38.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:22:38.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:38.379 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:38.379 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: Reconfiguring crash.vm08 (monmap changed)... 2026-03-10T10:22:38.379 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm08", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-10T10:22:38.379 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:38.379 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: Reconfiguring daemon crash.vm08 on vm08 2026-03-10T10:22:38.379 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:38.379 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm08.rwpjaf", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:22:38.379 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:22:38.379 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:38.379 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:38.379 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:22:38.379 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:22:38.379 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:38.379 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: from='osd.0 [v2:192.168.123.101:6802/3587270067,v1:192.168.123.101:6803/3587270067]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T10:22:38.380 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:38.380 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-10T10:22:38.380 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://192.168.123.101:9093"}]: dispatch 2026-03-10T10:22:38.380 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:38.380 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-10T10:22:38.380 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://192.168.123.101:3000"}]: dispatch 2026-03-10T10:22:38.380 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:38.380 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-10T10:22:38.380 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://192.168.123.101:9095"}]: dispatch 2026-03-10T10:22:38.380 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:38.380 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:22:38.380 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:38.380 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:22:38.380 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:22:38.380 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:22:38.380 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:39.016 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 10:22:39 vm01 ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f-osd-0[58396]: 2026-03-10T10:22:39.011+0000 7fd5cbbfd700 -1 osd.0 0 waiting for initial osdmap 2026-03-10T10:22:39.310 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:39 vm01 ceph-mon[48597]: Reconfiguring mgr.vm08.rwpjaf (monmap changed)... 2026-03-10T10:22:39.310 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:39 vm01 ceph-mon[48597]: Reconfiguring daemon mgr.vm08.rwpjaf on vm08 2026-03-10T10:22:39.310 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:39 vm01 ceph-mon[48597]: Reconfiguring mon.vm08 (monmap changed)... 2026-03-10T10:22:39.310 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:39 vm01 ceph-mon[48597]: Reconfiguring daemon mon.vm08 on vm08 2026-03-10T10:22:39.310 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:39 vm01 ceph-mon[48597]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:22:39.310 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:39 vm01 ceph-mon[48597]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-10T10:22:39.310 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:39 vm01 ceph-mon[48597]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://192.168.123.101:9093"}]: dispatch 2026-03-10T10:22:39.310 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:39 vm01 ceph-mon[48597]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-10T10:22:39.310 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:39 vm01 ceph-mon[48597]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://192.168.123.101:3000"}]: dispatch 2026-03-10T10:22:39.310 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:39 vm01 ceph-mon[48597]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-10T10:22:39.310 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:39 vm01 ceph-mon[48597]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://192.168.123.101:9095"}]: dispatch 2026-03-10T10:22:39.310 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:39 vm01 ceph-mon[48597]: from='client.14264 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:22:39.310 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:39 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:39.310 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:39 vm01 ceph-mon[48597]: from='osd.0 [v2:192.168.123.101:6802/3587270067,v1:192.168.123.101:6803/3587270067]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-10T10:22:39.310 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:39 vm01 ceph-mon[48597]: osdmap e7: 1 total, 0 up, 1 in 2026-03-10T10:22:39.310 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:39 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:22:39.310 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:39 vm01 ceph-mon[48597]: from='osd.0 [v2:192.168.123.101:6802/3587270067,v1:192.168.123.101:6803/3587270067]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T10:22:39.310 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:39 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/1633487310' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "5fd2081b-4e93-4c56-8dde-73e86252e942"}]: dispatch 2026-03-10T10:22:39.310 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:39 vm01 ceph-mon[48597]: from='osd.0 [v2:192.168.123.101:6802/3587270067,v1:192.168.123.101:6803/3587270067]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-10T10:22:39.310 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:39 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/1633487310' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "5fd2081b-4e93-4c56-8dde-73e86252e942"}]': finished 2026-03-10T10:22:39.310 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:39 vm01 ceph-mon[48597]: osdmap e8: 2 total, 0 up, 2 in 2026-03-10T10:22:39.310 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:39 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:22:39.310 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:39 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:22:39.310 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:39 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:22:39.311 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 10:22:39 vm01 ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f-osd-0[58396]: 2026-03-10T10:22:39.029+0000 7fd5c5d93700 -1 osd.0 8 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T10:22:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:39 vm08 ceph-mon[50506]: Reconfiguring mgr.vm08.rwpjaf (monmap changed)... 2026-03-10T10:22:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:39 vm08 ceph-mon[50506]: Reconfiguring daemon mgr.vm08.rwpjaf on vm08 2026-03-10T10:22:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:39 vm08 ceph-mon[50506]: Reconfiguring mon.vm08 (monmap changed)... 2026-03-10T10:22:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:39 vm08 ceph-mon[50506]: Reconfiguring daemon mon.vm08 on vm08 2026-03-10T10:22:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:39 vm08 ceph-mon[50506]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:22:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:39 vm08 ceph-mon[50506]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-10T10:22:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:39 vm08 ceph-mon[50506]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://192.168.123.101:9093"}]: dispatch 2026-03-10T10:22:39.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:39 vm08 ceph-mon[50506]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-10T10:22:39.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:39 vm08 ceph-mon[50506]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://192.168.123.101:3000"}]: dispatch 2026-03-10T10:22:39.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:39 vm08 ceph-mon[50506]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-10T10:22:39.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:39 vm08 ceph-mon[50506]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://192.168.123.101:9095"}]: dispatch 2026-03-10T10:22:39.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:39 vm08 ceph-mon[50506]: from='client.14264 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:22:39.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:39 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:39.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:39 vm08 ceph-mon[50506]: from='osd.0 [v2:192.168.123.101:6802/3587270067,v1:192.168.123.101:6803/3587270067]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-10T10:22:39.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:39 vm08 ceph-mon[50506]: osdmap e7: 1 total, 0 up, 1 in 2026-03-10T10:22:39.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:39 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:22:39.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:39 vm08 ceph-mon[50506]: from='osd.0 [v2:192.168.123.101:6802/3587270067,v1:192.168.123.101:6803/3587270067]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T10:22:39.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:39 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/1633487310' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "5fd2081b-4e93-4c56-8dde-73e86252e942"}]: dispatch 2026-03-10T10:22:39.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:39 vm08 ceph-mon[50506]: from='osd.0 [v2:192.168.123.101:6802/3587270067,v1:192.168.123.101:6803/3587270067]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-10T10:22:39.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:39 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/1633487310' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "5fd2081b-4e93-4c56-8dde-73e86252e942"}]': finished 2026-03-10T10:22:39.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:39 vm08 ceph-mon[50506]: osdmap e8: 2 total, 0 up, 2 in 2026-03-10T10:22:39.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:39 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:22:39.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:39 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:22:39.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:39 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:22:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:40 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:40 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/1499126561' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:22:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:40 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:22:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:40 vm08 ceph-mon[50506]: osd.0 [v2:192.168.123.101:6802/3587270067,v1:192.168.123.101:6803/3587270067] boot 2026-03-10T10:22:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:40 vm08 ceph-mon[50506]: osdmap e9: 2 total, 1 up, 2 in 2026-03-10T10:22:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:40 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:22:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:40 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:22:40.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:40 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:40.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:40 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/1499126561' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:22:40.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:40 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:22:40.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:40 vm01 ceph-mon[48597]: osd.0 [v2:192.168.123.101:6802/3587270067,v1:192.168.123.101:6803/3587270067] boot 2026-03-10T10:22:40.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:40 vm01 ceph-mon[48597]: osdmap e9: 2 total, 1 up, 2 in 2026-03-10T10:22:40.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:40 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:22:40.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:40 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:22:41.763 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:41 vm01 ceph-mon[48597]: purged_snaps scrub starts 2026-03-10T10:22:41.763 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:41 vm01 ceph-mon[48597]: purged_snaps scrub ok 2026-03-10T10:22:41.763 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:41 vm01 ceph-mon[48597]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:22:41.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:41 vm08 ceph-mon[50506]: purged_snaps scrub starts 2026-03-10T10:22:41.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:41 vm08 ceph-mon[50506]: purged_snaps scrub ok 2026-03-10T10:22:41.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:41 vm08 ceph-mon[50506]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:22:42.728 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:42 vm01 ceph-mon[48597]: osdmap e10: 2 total, 1 up, 2 in 2026-03-10T10:22:42.728 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:42 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:22:42.728 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:42 vm01 ceph-mon[48597]: pgmap v14: 0 pgs: ; 0 B data, 4.8 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:22:42.740 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:42 vm08 ceph-mon[50506]: osdmap e10: 2 total, 1 up, 2 in 2026-03-10T10:22:42.740 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:42 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:22:42.740 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:42 vm08 ceph-mon[50506]: pgmap v14: 0 pgs: ; 0 B data, 4.8 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:22:43.861 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:43 vm01 ceph-mon[48597]: Detected new or changed devices on vm01 2026-03-10T10:22:43.861 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:43 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:43.861 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:43 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:22:43.861 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:43 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:43.861 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:43 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:22:43.861 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:43 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:43.861 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:43 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:22:43.861 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:43 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T10:22:43.861 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:43 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:43.861 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:43 vm01 ceph-mon[48597]: Deploying daemon osd.1 on vm01 2026-03-10T10:22:43.861 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:43 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:44.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:43 vm08 ceph-mon[50506]: Detected new or changed devices on vm01 2026-03-10T10:22:44.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:43 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:44.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:43 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:22:44.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:43 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:44.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:43 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:22:44.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:43 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:44.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:43 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:22:44.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:43 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T10:22:44.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:43 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:44.095 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:43 vm08 ceph-mon[50506]: Deploying daemon osd.1 on vm01 2026-03-10T10:22:44.095 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:43 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:45.278 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:45 vm01 ceph-mon[48597]: pgmap v15: 0 pgs: ; 0 B data, 4.8 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:22:45.279 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:45 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:45.279 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:45 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:45.279 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:45 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:45.279 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:45 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:22:45.279 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:45 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:45.279 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:45 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:22:45.313 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:45 vm08 ceph-mon[50506]: pgmap v15: 0 pgs: ; 0 B data, 4.8 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:22:45.313 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:45 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:45.313 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:45 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:45.313 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:45 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:45.313 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:45 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:22:45.313 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:45 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:45.313 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:45 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:22:45.899 INFO:teuthology.orchestra.run.vm01.stdout:Created osd(s) 1 on host 'vm01' 2026-03-10T10:22:45.970 DEBUG:teuthology.orchestra.run.vm01:osd.1> sudo journalctl -f -n 0 -u ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f@osd.1.service 2026-03-10T10:22:45.981 INFO:tasks.cephadm:Deploying osd.2 on vm01 with /dev/vdc... 2026-03-10T10:22:45.981 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- lvm zap /dev/vdc 2026-03-10T10:22:46.454 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 10 10:22:46 vm01 ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f-osd-1[61910]: 2026-03-10T10:22:46.338+0000 7ff46fd433c0 -1 osd.1 0 log_to_monitors true 2026-03-10T10:22:46.769 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:46 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:46.770 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:46 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:46.770 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:46 vm01 ceph-mon[48597]: pgmap v16: 0 pgs: ; 0 B data, 4.8 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:22:46.770 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:46 vm01 ceph-mon[48597]: from='osd.1 [v2:192.168.123.101:6810/3090715425,v1:192.168.123.101:6811/3090715425]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T10:22:46.770 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:46 vm01 ceph-mon[48597]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T10:22:46.770 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:46 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:46.829 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T10:22:46.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:46 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:46.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:46 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:46.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:46 vm08 ceph-mon[50506]: pgmap v16: 0 pgs: ; 0 B data, 4.8 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:22:46.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:46 vm08 ceph-mon[50506]: from='osd.1 [v2:192.168.123.101:6810/3090715425,v1:192.168.123.101:6811/3090715425]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T10:22:46.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:46 vm08 ceph-mon[50506]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T10:22:46.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:46 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:46.847 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph orch daemon add osd vm01:/dev/vdc 2026-03-10T10:22:47.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:47 vm01 ceph-mon[48597]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-10T10:22:47.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:47 vm01 ceph-mon[48597]: osdmap e11: 2 total, 1 up, 2 in 2026-03-10T10:22:47.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:47 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:22:47.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:47 vm01 ceph-mon[48597]: from='osd.1 [v2:192.168.123.101:6810/3090715425,v1:192.168.123.101:6811/3090715425]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T10:22:47.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:47 vm01 ceph-mon[48597]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T10:22:47.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:47 vm01 ceph-mon[48597]: from='client.14280 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:22:47.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:47 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:22:47.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:47 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:22:47.961 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:47 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:47.961 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:47 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:47.961 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:47 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:22:47.961 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:47 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:47.961 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:47 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:22:47.961 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:47 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:47.961 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:47 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:22:47.961 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 10 10:22:47 vm01 ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f-osd-1[61910]: 2026-03-10T10:22:47.912+0000 7ff466746700 -1 osd.1 0 waiting for initial osdmap 2026-03-10T10:22:47.961 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 10 10:22:47 vm01 ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f-osd-1[61910]: 2026-03-10T10:22:47.917+0000 7ff4630e1700 -1 osd.1 12 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T10:22:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:47 vm08 ceph-mon[50506]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-10T10:22:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:47 vm08 ceph-mon[50506]: osdmap e11: 2 total, 1 up, 2 in 2026-03-10T10:22:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:47 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:22:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:47 vm08 ceph-mon[50506]: from='osd.1 [v2:192.168.123.101:6810/3090715425,v1:192.168.123.101:6811/3090715425]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T10:22:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:47 vm08 ceph-mon[50506]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T10:22:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:47 vm08 ceph-mon[50506]: from='client.14280 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:22:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:47 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:22:48.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:47 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:22:48.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:47 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:48.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:47 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:48.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:47 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:22:48.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:47 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:48.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:47 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:22:48.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:47 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:48.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:47 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:22:49.119 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:48 vm01 ceph-mon[48597]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-10T10:22:49.119 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:48 vm01 ceph-mon[48597]: osdmap e12: 2 total, 1 up, 2 in 2026-03-10T10:22:49.119 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:48 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:22:49.119 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:48 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:22:49.119 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:48 vm01 ceph-mon[48597]: pgmap v19: 0 pgs: ; 0 B data, 4.8 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:22:49.119 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:48 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/306628857' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "7b14edda-e6e2-4ff6-aa30-d875c9b27b49"}]: dispatch 2026-03-10T10:22:49.119 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:48 vm01 ceph-mon[48597]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "7b14edda-e6e2-4ff6-aa30-d875c9b27b49"}]: dispatch 2026-03-10T10:22:49.119 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:48 vm01 ceph-mon[48597]: osd.1 [v2:192.168.123.101:6810/3090715425,v1:192.168.123.101:6811/3090715425] boot 2026-03-10T10:22:49.119 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:48 vm01 ceph-mon[48597]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "7b14edda-e6e2-4ff6-aa30-d875c9b27b49"}]': finished 2026-03-10T10:22:49.119 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:48 vm01 ceph-mon[48597]: osdmap e13: 3 total, 2 up, 3 in 2026-03-10T10:22:49.119 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:48 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:22:49.119 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:48 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:22:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:48 vm08 ceph-mon[50506]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-10T10:22:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:48 vm08 ceph-mon[50506]: osdmap e12: 2 total, 1 up, 2 in 2026-03-10T10:22:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:48 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:22:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:48 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:22:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:48 vm08 ceph-mon[50506]: pgmap v19: 0 pgs: ; 0 B data, 4.8 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:22:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:48 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/306628857' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "7b14edda-e6e2-4ff6-aa30-d875c9b27b49"}]: dispatch 2026-03-10T10:22:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:48 vm08 ceph-mon[50506]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "7b14edda-e6e2-4ff6-aa30-d875c9b27b49"}]: dispatch 2026-03-10T10:22:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:48 vm08 ceph-mon[50506]: osd.1 [v2:192.168.123.101:6810/3090715425,v1:192.168.123.101:6811/3090715425] boot 2026-03-10T10:22:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:48 vm08 ceph-mon[50506]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "7b14edda-e6e2-4ff6-aa30-d875c9b27b49"}]': finished 2026-03-10T10:22:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:48 vm08 ceph-mon[50506]: osdmap e13: 3 total, 2 up, 3 in 2026-03-10T10:22:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:48 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:22:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:48 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:22:50.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:49 vm08 ceph-mon[50506]: purged_snaps scrub starts 2026-03-10T10:22:50.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:49 vm08 ceph-mon[50506]: purged_snaps scrub ok 2026-03-10T10:22:50.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:49 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/27930256' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:22:50.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:49 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:50.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:49 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:50.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:49 vm01 ceph-mon[48597]: purged_snaps scrub starts 2026-03-10T10:22:50.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:49 vm01 ceph-mon[48597]: purged_snaps scrub ok 2026-03-10T10:22:50.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:49 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/27930256' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:22:50.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:49 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:50.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:49 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:51.527 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:51 vm01 ceph-mon[48597]: pgmap v21: 0 pgs: ; 0 B data, 9.6 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:22:51.527 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:51 vm01 ceph-mon[48597]: osdmap e14: 3 total, 2 up, 3 in 2026-03-10T10:22:51.527 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:51 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:22:51.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:51 vm08 ceph-mon[50506]: pgmap v21: 0 pgs: ; 0 B data, 9.6 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:22:51.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:51 vm08 ceph-mon[50506]: osdmap e14: 3 total, 2 up, 3 in 2026-03-10T10:22:51.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:51 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:22:52.409 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:52 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T10:22:52.409 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:52 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:52.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:52 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T10:22:52.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:52 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:53.279 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:53 vm01 ceph-mon[48597]: Deploying daemon osd.2 on vm01 2026-03-10T10:22:53.279 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:53 vm01 ceph-mon[48597]: pgmap v23: 0 pgs: ; 0 B data, 9.7 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:22:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:53 vm08 ceph-mon[50506]: Deploying daemon osd.2 on vm01 2026-03-10T10:22:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:53 vm08 ceph-mon[50506]: pgmap v23: 0 pgs: ; 0 B data, 9.7 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:22:54.472 INFO:teuthology.orchestra.run.vm01.stdout:Created osd(s) 2 on host 'vm01' 2026-03-10T10:22:54.528 DEBUG:teuthology.orchestra.run.vm01:osd.2> sudo journalctl -f -n 0 -u ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f@osd.2.service 2026-03-10T10:22:54.529 INFO:tasks.cephadm:Deploying osd.3 on vm01 with /dev/vdb... 2026-03-10T10:22:54.529 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- lvm zap /dev/vdb 2026-03-10T10:22:54.737 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:54 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:54.737 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:54 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:22:54.738 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:54 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:54.738 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:54 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:22:54.738 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:54 vm01 ceph-mon[48597]: pgmap v24: 0 pgs: ; 0 B data, 9.8 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:22:54.738 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:54 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:54.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:54 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:54.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:54 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:22:54.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:54 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:54.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:54 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:22:54.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:54 vm08 ceph-mon[50506]: pgmap v24: 0 pgs: ; 0 B data, 9.8 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:22:54.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:54 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:55.306 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T10:22:55.320 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph orch daemon add osd vm01:/dev/vdb 2026-03-10T10:22:55.771 INFO:journalctl@ceph.osd.2.vm01.stdout:Mar 10 10:22:55 vm01 ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f-osd-2[65216]: 2026-03-10T10:22:55.704+0000 7fef6ca933c0 -1 osd.2 0 log_to_monitors true 2026-03-10T10:22:56.451 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:56 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:56.451 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:56 vm01 ceph-mon[48597]: from='osd.2 [v2:192.168.123.101:6818/1447659236,v1:192.168.123.101:6819/1447659236]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T10:22:56.451 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:56 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:22:56.452 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:56 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:22:56.452 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:56 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:56.478 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:56 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:56.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:56 vm08 ceph-mon[50506]: from='osd.2 [v2:192.168.123.101:6818/1447659236,v1:192.168.123.101:6819/1447659236]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T10:22:56.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:56 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:22:56.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:56 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:22:56.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:56 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:57.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:57 vm01 ceph-mon[48597]: from='client.14296 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vdb", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:22:57.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:57 vm01 ceph-mon[48597]: pgmap v25: 0 pgs: ; 0 B data, 9.8 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:22:57.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:57 vm01 ceph-mon[48597]: from='osd.2 [v2:192.168.123.101:6818/1447659236,v1:192.168.123.101:6819/1447659236]' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-10T10:22:57.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:57 vm01 ceph-mon[48597]: osdmap e15: 3 total, 2 up, 3 in 2026-03-10T10:22:57.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:57 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:22:57.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:57 vm01 ceph-mon[48597]: from='osd.2 [v2:192.168.123.101:6818/1447659236,v1:192.168.123.101:6819/1447659236]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T10:22:57.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:57 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:57.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:57 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:22:57.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:57 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:57.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:57 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:22:57.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:57 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:57.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:57 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:22:57.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:57 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:57.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:57 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/3774770827' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "af0a15a7-05e5-4909-8944-efe4cc34a49e"}]: dispatch 2026-03-10T10:22:57.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:57 vm01 ceph-mon[48597]: from='osd.2 [v2:192.168.123.101:6818/1447659236,v1:192.168.123.101:6819/1447659236]' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-10T10:22:57.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:57 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/3774770827' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "af0a15a7-05e5-4909-8944-efe4cc34a49e"}]': finished 2026-03-10T10:22:57.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:57 vm01 ceph-mon[48597]: osdmap e16: 4 total, 2 up, 4 in 2026-03-10T10:22:57.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:57 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:22:57.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:57 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:22:57.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:57 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:22:57.369 INFO:journalctl@ceph.osd.2.vm01.stdout:Mar 10 10:22:56 vm01 ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f-osd-2[65216]: 2026-03-10T10:22:56.952+0000 7fef63496700 -1 osd.2 0 waiting for initial osdmap 2026-03-10T10:22:57.369 INFO:journalctl@ceph.osd.2.vm01.stdout:Mar 10 10:22:56 vm01 ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f-osd-2[65216]: 2026-03-10T10:22:56.967+0000 7fef5e62e700 -1 osd.2 16 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T10:22:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:57 vm08 ceph-mon[50506]: from='client.14296 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vdb", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:22:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:57 vm08 ceph-mon[50506]: pgmap v25: 0 pgs: ; 0 B data, 9.8 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:22:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:57 vm08 ceph-mon[50506]: from='osd.2 [v2:192.168.123.101:6818/1447659236,v1:192.168.123.101:6819/1447659236]' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-10T10:22:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:57 vm08 ceph-mon[50506]: osdmap e15: 3 total, 2 up, 3 in 2026-03-10T10:22:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:57 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:22:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:57 vm08 ceph-mon[50506]: from='osd.2 [v2:192.168.123.101:6818/1447659236,v1:192.168.123.101:6819/1447659236]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T10:22:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:57 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:57 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:22:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:57 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:57 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:22:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:57 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:22:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:57 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:22:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:57 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:22:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:57 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/3774770827' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "af0a15a7-05e5-4909-8944-efe4cc34a49e"}]: dispatch 2026-03-10T10:22:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:57 vm08 ceph-mon[50506]: from='osd.2 [v2:192.168.123.101:6818/1447659236,v1:192.168.123.101:6819/1447659236]' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-10T10:22:57.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:57 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/3774770827' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "af0a15a7-05e5-4909-8944-efe4cc34a49e"}]': finished 2026-03-10T10:22:57.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:57 vm08 ceph-mon[50506]: osdmap e16: 4 total, 2 up, 4 in 2026-03-10T10:22:57.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:57 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:22:57.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:57 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:22:57.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:57 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:22:58.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:58 vm08 ceph-mon[50506]: Detected new or changed devices on vm01 2026-03-10T10:22:58.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:58 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/1461705168' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:22:58.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:58 vm08 ceph-mon[50506]: osd.2 [v2:192.168.123.101:6818/1447659236,v1:192.168.123.101:6819/1447659236] boot 2026-03-10T10:22:58.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:58 vm08 ceph-mon[50506]: osdmap e17: 4 total, 3 up, 4 in 2026-03-10T10:22:58.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:58 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:22:58.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:58 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:22:58.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:58 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32}]: dispatch 2026-03-10T10:22:58.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:58 vm01 ceph-mon[48597]: Detected new or changed devices on vm01 2026-03-10T10:22:58.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:58 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/1461705168' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:22:58.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:58 vm01 ceph-mon[48597]: osd.2 [v2:192.168.123.101:6818/1447659236,v1:192.168.123.101:6819/1447659236] boot 2026-03-10T10:22:58.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:58 vm01 ceph-mon[48597]: osdmap e17: 4 total, 3 up, 4 in 2026-03-10T10:22:58.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:58 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:22:58.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:58 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:22:58.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:58 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32}]: dispatch 2026-03-10T10:22:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:59 vm08 ceph-mon[50506]: purged_snaps scrub starts 2026-03-10T10:22:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:59 vm08 ceph-mon[50506]: purged_snaps scrub ok 2026-03-10T10:22:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:59 vm08 ceph-mon[50506]: pgmap v29: 0 pgs: ; 0 B data, 9.8 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:22:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:59 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32}]': finished 2026-03-10T10:22:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:59 vm08 ceph-mon[50506]: osdmap e18: 4 total, 3 up, 4 in 2026-03-10T10:22:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:59 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:22:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:22:59 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T10:22:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:59 vm01 ceph-mon[48597]: purged_snaps scrub starts 2026-03-10T10:22:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:59 vm01 ceph-mon[48597]: purged_snaps scrub ok 2026-03-10T10:22:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:59 vm01 ceph-mon[48597]: pgmap v29: 0 pgs: ; 0 B data, 9.8 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:22:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:59 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32}]': finished 2026-03-10T10:22:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:59 vm01 ceph-mon[48597]: osdmap e18: 4 total, 3 up, 4 in 2026-03-10T10:22:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:59 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:22:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:22:59 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T10:23:01.218 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:01 vm01 ceph-mon[48597]: pgmap v31: 1 pgs: 1 unknown; 0 B data, 15 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:23:01.219 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:01 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-10T10:23:01.219 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:01 vm01 ceph-mon[48597]: osdmap e19: 4 total, 3 up, 4 in 2026-03-10T10:23:01.219 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:01 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:23:01.219 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:01 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-10T10:23:01.219 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:01 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:01.488 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 10:23:01 vm01 sudo[67940]: ceph : TTY=unknown ; PWD=/ ; USER=root ; COMMAND=/sbin/smartctl -x --json=o /dev/vde 2026-03-10T10:23:01.488 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 10:23:01 vm01 sudo[67940]: pam_unix(sudo:session): session opened for user root by (uid=0) 2026-03-10T10:23:01.488 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 10:23:01 vm01 sudo[67940]: pam_unix(sudo:session): session closed for user root 2026-03-10T10:23:01.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:01 vm08 ceph-mon[50506]: pgmap v31: 1 pgs: 1 unknown; 0 B data, 15 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:23:01.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:01 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-10T10:23:01.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:01 vm08 ceph-mon[50506]: osdmap e19: 4 total, 3 up, 4 in 2026-03-10T10:23:01.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:01 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:23:01.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:01 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-10T10:23:01.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:01 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:01.752 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 10 10:23:01 vm01 sudo[67998]: ceph : TTY=unknown ; PWD=/ ; USER=root ; COMMAND=/sbin/smartctl -x --json=o /dev/vdd 2026-03-10T10:23:01.752 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 10 10:23:01 vm01 sudo[67998]: pam_unix(sudo:session): session opened for user root by (uid=0) 2026-03-10T10:23:01.752 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 10 10:23:01 vm01 sudo[67998]: pam_unix(sudo:session): session closed for user root 2026-03-10T10:23:02.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:02 vm01 sudo[68095]: ceph : TTY=unknown ; PWD=/ ; USER=root ; COMMAND=/sbin/smartctl -x --json=o /dev/vda 2026-03-10T10:23:02.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:02 vm01 sudo[68095]: pam_unix(sudo:session): session opened for user root by (uid=0) 2026-03-10T10:23:02.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:02 vm01 sudo[68095]: pam_unix(sudo:session): session closed for user root 2026-03-10T10:23:02.118 INFO:journalctl@ceph.osd.2.vm01.stdout:Mar 10 10:23:01 vm01 sudo[68042]: ceph : TTY=unknown ; PWD=/ ; USER=root ; COMMAND=/sbin/smartctl -x --json=o /dev/vdc 2026-03-10T10:23:02.118 INFO:journalctl@ceph.osd.2.vm01.stdout:Mar 10 10:23:01 vm01 sudo[68042]: pam_unix(sudo:session): session opened for user root by (uid=0) 2026-03-10T10:23:02.118 INFO:journalctl@ceph.osd.2.vm01.stdout:Mar 10 10:23:01 vm01 sudo[68042]: pam_unix(sudo:session): session closed for user root 2026-03-10T10:23:02.530 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:02 vm01 ceph-mon[48597]: Deploying daemon osd.3 on vm01 2026-03-10T10:23:02.530 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:02 vm01 ceph-mon[48597]: osdmap e20: 4 total, 3 up, 4 in 2026-03-10T10:23:02.530 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:02 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:23:02.530 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:02 vm01 ceph-mon[48597]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T10:23:02.530 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:02 vm01 ceph-mon[48597]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T10:23:02.530 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:02 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-10T10:23:02.530 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:02 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-10T10:23:02.530 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:02 vm01 ceph-mon[48597]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T10:23:02.530 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:02 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-10T10:23:02.530 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:02 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-10T10:23:02.530 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:02 vm01 ceph-mon[48597]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T10:23:02.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:02 vm08 sudo[52719]: ceph : TTY=unknown ; PWD=/ ; USER=root ; COMMAND=/sbin/smartctl -x --json=o /dev/vda 2026-03-10T10:23:02.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:02 vm08 sudo[52719]: pam_unix(sudo:session): session opened for user root by (uid=0) 2026-03-10T10:23:02.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:02 vm08 sudo[52719]: pam_unix(sudo:session): session closed for user root 2026-03-10T10:23:02.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:02 vm08 ceph-mon[50506]: Deploying daemon osd.3 on vm01 2026-03-10T10:23:02.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:02 vm08 ceph-mon[50506]: osdmap e20: 4 total, 3 up, 4 in 2026-03-10T10:23:02.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:02 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:23:02.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:02 vm08 ceph-mon[50506]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T10:23:02.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:02 vm08 ceph-mon[50506]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T10:23:02.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:02 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-10T10:23:02.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:02 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-10T10:23:02.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:02 vm08 ceph-mon[50506]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T10:23:02.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:02 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-10T10:23:02.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:02 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-10T10:23:02.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:02 vm08 ceph-mon[50506]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T10:23:03.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:03 vm01 ceph-mon[48597]: pgmap v34: 1 pgs: 1 unknown; 0 B data, 15 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:23:03.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:03 vm01 ceph-mon[48597]: mgrmap e19: vm01.nkkzmc(active, since 52s), standbys: vm08.rwpjaf 2026-03-10T10:23:03.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:03 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:03.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:03 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:23:03.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:03 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:03.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:03 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:23:03.493 INFO:teuthology.orchestra.run.vm01.stdout:Created osd(s) 3 on host 'vm01' 2026-03-10T10:23:03.572 DEBUG:teuthology.orchestra.run.vm01:osd.3> sudo journalctl -f -n 0 -u ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f@osd.3.service 2026-03-10T10:23:03.574 INFO:tasks.cephadm:Deploying osd.4 on vm08 with /dev/vde... 2026-03-10T10:23:03.574 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- lvm zap /dev/vde 2026-03-10T10:23:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:03 vm08 ceph-mon[50506]: pgmap v34: 1 pgs: 1 unknown; 0 B data, 15 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:23:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:03 vm08 ceph-mon[50506]: mgrmap e19: vm01.nkkzmc(active, since 52s), standbys: vm08.rwpjaf 2026-03-10T10:23:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:03 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:03 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:23:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:03 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:03 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:23:04.182 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-10T10:23:04.195 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph orch daemon add osd vm08:/dev/vde 2026-03-10T10:23:04.487 INFO:journalctl@ceph.osd.3.vm01.stdout:Mar 10 10:23:04 vm01 ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f-osd-3[68201]: 2026-03-10T10:23:04.418+0000 7ff9082933c0 -1 osd.3 0 log_to_monitors true 2026-03-10T10:23:04.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:04 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:04.725 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:04 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:04.726 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:04 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:04.726 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:04 vm08 ceph-mon[50506]: pgmap v35: 1 pgs: 1 active+clean; 449 KiB data, 17 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:23:04.726 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:04 vm08 ceph-mon[50506]: from='osd.3 [v2:192.168.123.101:6826/3038741761,v1:192.168.123.101:6827/3038741761]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-10T10:23:04.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:04 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:04.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:04 vm01 ceph-mon[48597]: pgmap v35: 1 pgs: 1 active+clean; 449 KiB data, 17 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:23:04.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:04 vm01 ceph-mon[48597]: from='osd.3 [v2:192.168.123.101:6826/3038741761,v1:192.168.123.101:6827/3038741761]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-10T10:23:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:05 vm08 ceph-mon[50506]: from='client.24151 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm08:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:23:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:05 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:23:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:05 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:23:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:05 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:05 vm08 ceph-mon[50506]: from='osd.3 [v2:192.168.123.101:6826/3038741761,v1:192.168.123.101:6827/3038741761]' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-10T10:23:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:05 vm08 ceph-mon[50506]: osdmap e21: 4 total, 3 up, 4 in 2026-03-10T10:23:05.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:05 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:23:05.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:05 vm08 ceph-mon[50506]: from='osd.3 [v2:192.168.123.101:6826/3038741761,v1:192.168.123.101:6827/3038741761]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T10:23:05.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:05 vm08 ceph-mon[50506]: Detected new or changed devices on vm01 2026-03-10T10:23:05.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:05 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:05.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:05 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:23:05.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:05 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:05.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:05 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:23:05.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:05 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:05.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:05 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:23:05.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:05 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:05.868 INFO:journalctl@ceph.osd.3.vm01.stdout:Mar 10 10:23:05 vm01 ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f-osd-3[68201]: 2026-03-10T10:23:05.686+0000 7ff8fec96700 -1 osd.3 0 waiting for initial osdmap 2026-03-10T10:23:05.868 INFO:journalctl@ceph.osd.3.vm01.stdout:Mar 10 10:23:05 vm01 ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f-osd-3[68201]: 2026-03-10T10:23:05.693+0000 7ff8f8e2c700 -1 osd.3 22 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T10:23:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:05 vm01 ceph-mon[48597]: from='client.24151 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm08:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:23:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:05 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:23:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:05 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:23:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:05 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:05 vm01 ceph-mon[48597]: from='osd.3 [v2:192.168.123.101:6826/3038741761,v1:192.168.123.101:6827/3038741761]' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-10T10:23:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:05 vm01 ceph-mon[48597]: osdmap e21: 4 total, 3 up, 4 in 2026-03-10T10:23:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:05 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:23:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:05 vm01 ceph-mon[48597]: from='osd.3 [v2:192.168.123.101:6826/3038741761,v1:192.168.123.101:6827/3038741761]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T10:23:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:05 vm01 ceph-mon[48597]: Detected new or changed devices on vm01 2026-03-10T10:23:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:05 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:05 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:23:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:05 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:05 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:23:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:05 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:05 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:23:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:05 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:06.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:06 vm08 ceph-mon[50506]: from='client.? 192.168.123.108:0/2908509228' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "b7907d54-5467-4a9f-ae4f-614fa10c8000"}]: dispatch 2026-03-10T10:23:06.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:06 vm08 ceph-mon[50506]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "b7907d54-5467-4a9f-ae4f-614fa10c8000"}]: dispatch 2026-03-10T10:23:06.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:06 vm08 ceph-mon[50506]: from='osd.3 [v2:192.168.123.101:6826/3038741761,v1:192.168.123.101:6827/3038741761]' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-10T10:23:06.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:06 vm08 ceph-mon[50506]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "b7907d54-5467-4a9f-ae4f-614fa10c8000"}]': finished 2026-03-10T10:23:06.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:06 vm08 ceph-mon[50506]: osdmap e22: 5 total, 3 up, 5 in 2026-03-10T10:23:06.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:06 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:23:06.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:06 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:23:06.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:06 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:23:06.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:06 vm08 ceph-mon[50506]: pgmap v38: 1 pgs: 1 active+clean; 449 KiB data, 17 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:23:06.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:06 vm08 ceph-mon[50506]: from='client.? 192.168.123.108:0/3230869738' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:23:06.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:06 vm01 ceph-mon[48597]: from='client.? 192.168.123.108:0/2908509228' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "b7907d54-5467-4a9f-ae4f-614fa10c8000"}]: dispatch 2026-03-10T10:23:06.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:06 vm01 ceph-mon[48597]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "b7907d54-5467-4a9f-ae4f-614fa10c8000"}]: dispatch 2026-03-10T10:23:06.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:06 vm01 ceph-mon[48597]: from='osd.3 [v2:192.168.123.101:6826/3038741761,v1:192.168.123.101:6827/3038741761]' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-10T10:23:06.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:06 vm01 ceph-mon[48597]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "b7907d54-5467-4a9f-ae4f-614fa10c8000"}]': finished 2026-03-10T10:23:06.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:06 vm01 ceph-mon[48597]: osdmap e22: 5 total, 3 up, 5 in 2026-03-10T10:23:06.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:06 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:23:06.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:06 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:23:06.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:06 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:23:06.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:06 vm01 ceph-mon[48597]: pgmap v38: 1 pgs: 1 active+clean; 449 KiB data, 17 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:23:06.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:06 vm01 ceph-mon[48597]: from='client.? 192.168.123.108:0/3230869738' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:23:08.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:07 vm08 ceph-mon[50506]: purged_snaps scrub starts 2026-03-10T10:23:08.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:07 vm08 ceph-mon[50506]: purged_snaps scrub ok 2026-03-10T10:23:08.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:07 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:23:08.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:07 vm08 ceph-mon[50506]: osd.3 [v2:192.168.123.101:6826/3038741761,v1:192.168.123.101:6827/3038741761] boot 2026-03-10T10:23:08.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:07 vm08 ceph-mon[50506]: osdmap e23: 5 total, 4 up, 5 in 2026-03-10T10:23:08.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:07 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:23:08.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:07 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:23:08.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:07 vm01 ceph-mon[48597]: purged_snaps scrub starts 2026-03-10T10:23:08.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:07 vm01 ceph-mon[48597]: purged_snaps scrub ok 2026-03-10T10:23:08.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:07 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:23:08.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:07 vm01 ceph-mon[48597]: osd.3 [v2:192.168.123.101:6826/3038741761,v1:192.168.123.101:6827/3038741761] boot 2026-03-10T10:23:08.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:07 vm01 ceph-mon[48597]: osdmap e23: 5 total, 4 up, 5 in 2026-03-10T10:23:08.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:07 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:23:08.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:07 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:23:08.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:08 vm08 ceph-mon[50506]: osdmap e24: 5 total, 4 up, 5 in 2026-03-10T10:23:08.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:08 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:23:08.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:08 vm08 ceph-mon[50506]: pgmap v41: 1 pgs: 1 active+clean; 449 KiB data, 22 MiB used, 80 GiB / 80 GiB avail 2026-03-10T10:23:09.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:08 vm01 ceph-mon[48597]: osdmap e24: 5 total, 4 up, 5 in 2026-03-10T10:23:09.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:08 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:23:09.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:08 vm01 ceph-mon[48597]: pgmap v41: 1 pgs: 1 active+clean; 449 KiB data, 22 MiB used, 80 GiB / 80 GiB avail 2026-03-10T10:23:10.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-10T10:23:10.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:10.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:10 vm08 ceph-mon[50506]: Deploying daemon osd.4 on vm08 2026-03-10T10:23:10.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-10T10:23:10.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:10.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:10 vm01 ceph-mon[48597]: Deploying daemon osd.4 on vm08 2026-03-10T10:23:11.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:11 vm08 ceph-mon[50506]: pgmap v42: 1 pgs: 1 active+clean; 449 KiB data, 23 MiB used, 80 GiB / 80 GiB avail 2026-03-10T10:23:11.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:11 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:23:11.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:11 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:23:11.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:11 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:11.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:11 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:23:11.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:11 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:11.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:11 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:23:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:11 vm01 ceph-mon[48597]: pgmap v42: 1 pgs: 1 active+clean; 449 KiB data, 23 MiB used, 80 GiB / 80 GiB avail 2026-03-10T10:23:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:23:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:23:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:23:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:23:11.426 INFO:teuthology.orchestra.run.vm08.stdout:Created osd(s) 4 on host 'vm08' 2026-03-10T10:23:11.497 DEBUG:teuthology.orchestra.run.vm08:osd.4> sudo journalctl -f -n 0 -u ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f@osd.4.service 2026-03-10T10:23:11.498 INFO:tasks.cephadm:Deploying osd.5 on vm08 with /dev/vdd... 2026-03-10T10:23:11.498 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- lvm zap /dev/vdd 2026-03-10T10:23:12.178 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-10T10:23:12.192 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph orch daemon add osd vm08:/dev/vdd 2026-03-10T10:23:12.331 INFO:journalctl@ceph.osd.4.vm08.stdout:Mar 10 10:23:12 vm08 ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f-osd-4[54464]: 2026-03-10T10:23:12.054+0000 7ff6fddc03c0 -1 osd.4 0 log_to_monitors true 2026-03-10T10:23:12.589 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:12 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:12.589 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:12 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:12.589 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:12 vm08 ceph-mon[50506]: from='osd.4 [v2:192.168.123.108:6800/3235322654,v1:192.168.123.108:6801/3235322654]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-10T10:23:12.589 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:12 vm08 ceph-mon[50506]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-10T10:23:12.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:12 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:12.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:12 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:12.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:12 vm01 ceph-mon[48597]: from='osd.4 [v2:192.168.123.108:6800/3235322654,v1:192.168.123.108:6801/3235322654]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-10T10:23:12.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:12 vm01 ceph-mon[48597]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-10T10:23:13.438 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:13 vm08 ceph-mon[50506]: pgmap v43: 1 pgs: 1 active+clean; 449 KiB data, 23 MiB used, 80 GiB / 80 GiB avail 2026-03-10T10:23:13.438 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:13 vm08 ceph-mon[50506]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-10T10:23:13.438 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:13 vm08 ceph-mon[50506]: osdmap e25: 5 total, 4 up, 5 in 2026-03-10T10:23:13.438 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:13 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:23:13.438 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:13 vm08 ceph-mon[50506]: from='osd.4 [v2:192.168.123.108:6800/3235322654,v1:192.168.123.108:6801/3235322654]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-10T10:23:13.438 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:13 vm08 ceph-mon[50506]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-10T10:23:13.438 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:13 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:23:13.438 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:13 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:23:13.438 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:13 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:13.438 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:13 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:13.438 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:13 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:23:13.438 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:13 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:13.438 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:13 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:23:13.438 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:13 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:13.438 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:13 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:23:13.439 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:13 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:13.844 INFO:journalctl@ceph.osd.4.vm08.stdout:Mar 10 10:23:13 vm08 ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f-osd-4[54464]: 2026-03-10T10:23:13.510+0000 7ff6f5fc6700 -1 osd.4 0 waiting for initial osdmap 2026-03-10T10:23:13.844 INFO:journalctl@ceph.osd.4.vm08.stdout:Mar 10 10:23:13 vm08 ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f-osd-4[54464]: 2026-03-10T10:23:13.518+0000 7ff6ee959700 -1 osd.4 26 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T10:23:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:13 vm01 ceph-mon[48597]: pgmap v43: 1 pgs: 1 active+clean; 449 KiB data, 23 MiB used, 80 GiB / 80 GiB avail 2026-03-10T10:23:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:13 vm01 ceph-mon[48597]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-10T10:23:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:13 vm01 ceph-mon[48597]: osdmap e25: 5 total, 4 up, 5 in 2026-03-10T10:23:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:13 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:23:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:13 vm01 ceph-mon[48597]: from='osd.4 [v2:192.168.123.108:6800/3235322654,v1:192.168.123.108:6801/3235322654]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-10T10:23:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:13 vm01 ceph-mon[48597]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-10T10:23:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:13 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:23:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:13 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:23:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:13 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:13 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:13 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:23:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:13 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:13 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:23:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:13 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:13 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:23:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:13 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:14.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:14 vm08 ceph-mon[50506]: from='client.24169 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm08:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:23:14.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:14 vm08 ceph-mon[50506]: Detected new or changed devices on vm08 2026-03-10T10:23:14.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:14 vm08 ceph-mon[50506]: from='client.? 192.168.123.108:0/3787286349' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "3a21cd68-d40c-4c3b-a20b-efb92c0277a1"}]: dispatch 2026-03-10T10:23:14.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:14 vm08 ceph-mon[50506]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "3a21cd68-d40c-4c3b-a20b-efb92c0277a1"}]: dispatch 2026-03-10T10:23:14.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:14 vm08 ceph-mon[50506]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-10T10:23:14.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:14 vm08 ceph-mon[50506]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "3a21cd68-d40c-4c3b-a20b-efb92c0277a1"}]': finished 2026-03-10T10:23:14.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:14 vm08 ceph-mon[50506]: osdmap e26: 6 total, 4 up, 6 in 2026-03-10T10:23:14.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:14 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:23:14.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:14 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:23:14.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:14 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:23:14.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:14 vm08 ceph-mon[50506]: from='client.? 192.168.123.108:0/3007264458' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:23:14.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:14 vm08 ceph-mon[50506]: osd.4 [v2:192.168.123.108:6800/3235322654,v1:192.168.123.108:6801/3235322654] boot 2026-03-10T10:23:14.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:14 vm08 ceph-mon[50506]: osdmap e27: 6 total, 5 up, 6 in 2026-03-10T10:23:14.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:14 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:23:14.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:14 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:23:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:14 vm01 ceph-mon[48597]: from='client.24169 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm08:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:23:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:14 vm01 ceph-mon[48597]: Detected new or changed devices on vm08 2026-03-10T10:23:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:14 vm01 ceph-mon[48597]: from='client.? 192.168.123.108:0/3787286349' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "3a21cd68-d40c-4c3b-a20b-efb92c0277a1"}]: dispatch 2026-03-10T10:23:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:14 vm01 ceph-mon[48597]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "3a21cd68-d40c-4c3b-a20b-efb92c0277a1"}]: dispatch 2026-03-10T10:23:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:14 vm01 ceph-mon[48597]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-10T10:23:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:14 vm01 ceph-mon[48597]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "3a21cd68-d40c-4c3b-a20b-efb92c0277a1"}]': finished 2026-03-10T10:23:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:14 vm01 ceph-mon[48597]: osdmap e26: 6 total, 4 up, 6 in 2026-03-10T10:23:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:14 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:23:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:14 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:23:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:14 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:23:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:14 vm01 ceph-mon[48597]: from='client.? 192.168.123.108:0/3007264458' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:23:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:14 vm01 ceph-mon[48597]: osd.4 [v2:192.168.123.108:6800/3235322654,v1:192.168.123.108:6801/3235322654] boot 2026-03-10T10:23:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:14 vm01 ceph-mon[48597]: osdmap e27: 6 total, 5 up, 6 in 2026-03-10T10:23:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:14 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:23:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:14 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:23:15.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:15 vm08 ceph-mon[50506]: purged_snaps scrub starts 2026-03-10T10:23:15.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:15 vm08 ceph-mon[50506]: purged_snaps scrub ok 2026-03-10T10:23:15.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:15 vm08 ceph-mon[50506]: pgmap v47: 1 pgs: 1 peering; 449 KiB data, 23 MiB used, 80 GiB / 80 GiB avail 2026-03-10T10:23:15.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:15 vm08 ceph-mon[50506]: osdmap e28: 6 total, 5 up, 6 in 2026-03-10T10:23:15.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:15 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:23:15.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:15 vm01 ceph-mon[48597]: purged_snaps scrub starts 2026-03-10T10:23:15.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:15 vm01 ceph-mon[48597]: purged_snaps scrub ok 2026-03-10T10:23:15.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:15 vm01 ceph-mon[48597]: pgmap v47: 1 pgs: 1 peering; 449 KiB data, 23 MiB used, 80 GiB / 80 GiB avail 2026-03-10T10:23:15.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:15 vm01 ceph-mon[48597]: osdmap e28: 6 total, 5 up, 6 in 2026-03-10T10:23:15.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:15 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:23:17.239 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:17 vm08 ceph-mon[50506]: osdmap e29: 6 total, 5 up, 6 in 2026-03-10T10:23:17.240 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:17 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:23:17.240 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:17 vm08 ceph-mon[50506]: pgmap v50: 1 pgs: 1 peering; 449 KiB data, 28 MiB used, 100 GiB / 100 GiB avail 2026-03-10T10:23:17.240 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:17 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-10T10:23:17.240 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:17 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:17.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:17 vm01 ceph-mon[48597]: osdmap e29: 6 total, 5 up, 6 in 2026-03-10T10:23:17.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:17 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:23:17.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:17 vm01 ceph-mon[48597]: pgmap v50: 1 pgs: 1 peering; 449 KiB data, 28 MiB used, 100 GiB / 100 GiB avail 2026-03-10T10:23:17.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:17 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-10T10:23:17.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:17 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:18.298 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:18 vm08 ceph-mon[50506]: Deploying daemon osd.5 on vm08 2026-03-10T10:23:18.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:18 vm01 ceph-mon[48597]: Deploying daemon osd.5 on vm08 2026-03-10T10:23:18.974 INFO:teuthology.orchestra.run.vm08.stdout:Created osd(s) 5 on host 'vm08' 2026-03-10T10:23:19.033 DEBUG:teuthology.orchestra.run.vm08:osd.5> sudo journalctl -f -n 0 -u ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f@osd.5.service 2026-03-10T10:23:19.034 INFO:tasks.cephadm:Deploying osd.6 on vm08 with /dev/vdc... 2026-03-10T10:23:19.034 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- lvm zap /dev/vdc 2026-03-10T10:23:19.393 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:19 vm08 ceph-mon[50506]: pgmap v51: 1 pgs: 1 peering; 449 KiB data, 28 MiB used, 100 GiB / 100 GiB avail 2026-03-10T10:23:19.393 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:19 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:19.393 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:19 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:23:19.393 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:19 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:19.393 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:19 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:23:19.393 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:19 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:19.393 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:19 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:19 vm01 ceph-mon[48597]: pgmap v51: 1 pgs: 1 peering; 449 KiB data, 28 MiB used, 100 GiB / 100 GiB avail 2026-03-10T10:23:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:19 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:19 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:23:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:19 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:19 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:23:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:19 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:19 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:19.769 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-10T10:23:19.794 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph orch daemon add osd vm08:/dev/vdc 2026-03-10T10:23:20.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:20 vm08 ceph-mon[50506]: from='osd.5 [v2:192.168.123.108:6808/2436937416,v1:192.168.123.108:6809/2436937416]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-10T10:23:20.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:20 vm08 ceph-mon[50506]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-10T10:23:20.344 INFO:journalctl@ceph.osd.5.vm08.stdout:Mar 10 10:23:20 vm08 ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f-osd-5[57314]: 2026-03-10T10:23:20.082+0000 7f2da5c153c0 -1 osd.5 0 log_to_monitors true 2026-03-10T10:23:20.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:20 vm01 ceph-mon[48597]: from='osd.5 [v2:192.168.123.108:6808/2436937416,v1:192.168.123.108:6809/2436937416]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-10T10:23:20.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:20 vm01 ceph-mon[48597]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-10T10:23:21.092 INFO:journalctl@ceph.osd.5.vm08.stdout:Mar 10 10:23:21 vm08 ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f-osd-5[57314]: 2026-03-10T10:23:21.089+0000 7f2d9c618700 -1 osd.5 0 waiting for initial osdmap 2026-03-10T10:23:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:21 vm08 ceph-mon[50506]: pgmap v52: 1 pgs: 1 active+clean; 449 KiB data, 28 MiB used, 100 GiB / 100 GiB avail; 65 KiB/s, 0 objects/s recovering 2026-03-10T10:23:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:21 vm08 ceph-mon[50506]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-10T10:23:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:21 vm08 ceph-mon[50506]: osdmap e30: 6 total, 5 up, 6 in 2026-03-10T10:23:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:21 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:23:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:21 vm08 ceph-mon[50506]: from='osd.5 [v2:192.168.123.108:6808/2436937416,v1:192.168.123.108:6809/2436937416]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-10T10:23:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:21 vm08 ceph-mon[50506]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-10T10:23:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:21 vm08 ceph-mon[50506]: from='client.24187 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm08:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:23:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:21 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:23:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:21 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:23:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:21 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:21 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:21 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:23:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:21 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:21 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:23:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:21 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:21 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:23:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:21 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:21 vm08 ceph-mon[50506]: from='client.? 192.168.123.108:0/2278114564' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "162defaf-0c11-4032-b25b-1848a829ed6f"}]: dispatch 2026-03-10T10:23:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:21 vm08 ceph-mon[50506]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "162defaf-0c11-4032-b25b-1848a829ed6f"}]: dispatch 2026-03-10T10:23:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:21 vm08 ceph-mon[50506]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-10T10:23:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:21 vm08 ceph-mon[50506]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "162defaf-0c11-4032-b25b-1848a829ed6f"}]': finished 2026-03-10T10:23:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:21 vm08 ceph-mon[50506]: osdmap e31: 7 total, 5 up, 7 in 2026-03-10T10:23:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:21 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:23:21.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:21 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:23:21.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:21 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:23:21.345 INFO:journalctl@ceph.osd.5.vm08.stdout:Mar 10 10:23:21 vm08 ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f-osd-5[57314]: 2026-03-10T10:23:21.102+0000 7f2d987b2700 -1 osd.5 31 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T10:23:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:21 vm01 ceph-mon[48597]: pgmap v52: 1 pgs: 1 active+clean; 449 KiB data, 28 MiB used, 100 GiB / 100 GiB avail; 65 KiB/s, 0 objects/s recovering 2026-03-10T10:23:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:21 vm01 ceph-mon[48597]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-10T10:23:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:21 vm01 ceph-mon[48597]: osdmap e30: 6 total, 5 up, 6 in 2026-03-10T10:23:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:21 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:23:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:21 vm01 ceph-mon[48597]: from='osd.5 [v2:192.168.123.108:6808/2436937416,v1:192.168.123.108:6809/2436937416]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-10T10:23:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:21 vm01 ceph-mon[48597]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-10T10:23:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:21 vm01 ceph-mon[48597]: from='client.24187 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm08:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:23:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:21 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:23:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:21 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:23:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:21 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:21 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:21 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:23:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:21 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:21 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:23:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:21 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:21 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:23:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:21 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:21 vm01 ceph-mon[48597]: from='client.? 192.168.123.108:0/2278114564' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "162defaf-0c11-4032-b25b-1848a829ed6f"}]: dispatch 2026-03-10T10:23:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:21 vm01 ceph-mon[48597]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "162defaf-0c11-4032-b25b-1848a829ed6f"}]: dispatch 2026-03-10T10:23:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:21 vm01 ceph-mon[48597]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-10T10:23:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:21 vm01 ceph-mon[48597]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "162defaf-0c11-4032-b25b-1848a829ed6f"}]': finished 2026-03-10T10:23:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:21 vm01 ceph-mon[48597]: osdmap e31: 7 total, 5 up, 7 in 2026-03-10T10:23:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:21 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:23:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:21 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:23:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:21 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:23:22.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:22 vm08 ceph-mon[50506]: Detected new or changed devices on vm08 2026-03-10T10:23:22.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:22 vm08 ceph-mon[50506]: from='client.? 192.168.123.108:0/301736619' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:23:22.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:22 vm08 ceph-mon[50506]: osd.5 [v2:192.168.123.108:6808/2436937416,v1:192.168.123.108:6809/2436937416] boot 2026-03-10T10:23:22.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:22 vm08 ceph-mon[50506]: osdmap e32: 7 total, 6 up, 7 in 2026-03-10T10:23:22.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:22 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:23:22.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:22 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:23:22.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:22 vm01 ceph-mon[48597]: Detected new or changed devices on vm08 2026-03-10T10:23:22.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:22 vm01 ceph-mon[48597]: from='client.? 192.168.123.108:0/301736619' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:23:22.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:22 vm01 ceph-mon[48597]: osd.5 [v2:192.168.123.108:6808/2436937416,v1:192.168.123.108:6809/2436937416] boot 2026-03-10T10:23:22.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:22 vm01 ceph-mon[48597]: osdmap e32: 7 total, 6 up, 7 in 2026-03-10T10:23:22.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:22 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:23:22.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:22 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:23:23.507 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:23 vm01 ceph-mon[48597]: purged_snaps scrub starts 2026-03-10T10:23:23.507 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:23 vm01 ceph-mon[48597]: purged_snaps scrub ok 2026-03-10T10:23:23.507 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:23 vm01 ceph-mon[48597]: pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 28 MiB used, 100 GiB / 100 GiB avail; 65 KiB/s, 0 objects/s recovering 2026-03-10T10:23:23.507 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:23 vm01 ceph-mon[48597]: osdmap e33: 7 total, 6 up, 7 in 2026-03-10T10:23:23.507 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:23 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:23:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:23 vm08 ceph-mon[50506]: purged_snaps scrub starts 2026-03-10T10:23:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:23 vm08 ceph-mon[50506]: purged_snaps scrub ok 2026-03-10T10:23:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:23 vm08 ceph-mon[50506]: pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 28 MiB used, 100 GiB / 100 GiB avail; 65 KiB/s, 0 objects/s recovering 2026-03-10T10:23:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:23 vm08 ceph-mon[50506]: osdmap e33: 7 total, 6 up, 7 in 2026-03-10T10:23:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:23 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:23:25.280 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:25 vm08 ceph-mon[50506]: pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 34 MiB used, 120 GiB / 120 GiB avail 2026-03-10T10:23:25.281 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:25 vm08 ceph-mon[50506]: osdmap e34: 7 total, 6 up, 7 in 2026-03-10T10:23:25.281 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:25 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:23:25.281 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:25 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-10T10:23:25.281 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:25 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:25.281 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:25 vm08 ceph-mon[50506]: Deploying daemon osd.6 on vm08 2026-03-10T10:23:25.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:25 vm01 ceph-mon[48597]: pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 34 MiB used, 120 GiB / 120 GiB avail 2026-03-10T10:23:25.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:25 vm01 ceph-mon[48597]: osdmap e34: 7 total, 6 up, 7 in 2026-03-10T10:23:25.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:25 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:23:25.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:25 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-10T10:23:25.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:25 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:25.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:25 vm01 ceph-mon[48597]: Deploying daemon osd.6 on vm08 2026-03-10T10:23:27.282 INFO:teuthology.orchestra.run.vm08.stdout:Created osd(s) 6 on host 'vm08' 2026-03-10T10:23:27.364 DEBUG:teuthology.orchestra.run.vm08:osd.6> sudo journalctl -f -n 0 -u ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f@osd.6.service 2026-03-10T10:23:27.367 INFO:tasks.cephadm:Deploying osd.7 on vm08 with /dev/vdb... 2026-03-10T10:23:27.367 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- lvm zap /dev/vdb 2026-03-10T10:23:27.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:27 vm01 ceph-mon[48597]: pgmap v60: 1 pgs: 1 active+clean; 449 KiB data, 34 MiB used, 120 GiB / 120 GiB avail 2026-03-10T10:23:27.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:27 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:23:27.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:27 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:27.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:27 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:23:27.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:27 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:27.402 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:27 vm08 ceph-mon[50506]: pgmap v60: 1 pgs: 1 active+clean; 449 KiB data, 34 MiB used, 120 GiB / 120 GiB avail 2026-03-10T10:23:27.402 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:27 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:23:27.402 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:27 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:27.402 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:27 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:23:27.402 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:27 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:28.010 INFO:journalctl@ceph.osd.6.vm08.stdout:Mar 10 10:23:27 vm08 ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f-osd-6[60100]: 2026-03-10T10:23:27.833+0000 7fd6b2c383c0 -1 osd.6 0 log_to_monitors true 2026-03-10T10:23:28.287 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-10T10:23:28.306 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph orch daemon add osd vm08:/dev/vdb 2026-03-10T10:23:28.596 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:28 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:28.596 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:28 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:28.596 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:28 vm08 ceph-mon[50506]: from='osd.6 [v2:192.168.123.108:6816/3667011305,v1:192.168.123.108:6817/3667011305]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-10T10:23:28.596 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:28 vm08 ceph-mon[50506]: from='osd.6 ' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-10T10:23:28.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:28 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:28.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:28 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:28.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:28 vm01 ceph-mon[48597]: from='osd.6 [v2:192.168.123.108:6816/3667011305,v1:192.168.123.108:6817/3667011305]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-10T10:23:28.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:28 vm01 ceph-mon[48597]: from='osd.6 ' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-10T10:23:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:29 vm08 ceph-mon[50506]: pgmap v61: 1 pgs: 1 active+recovering; 449 KiB data, 35 MiB used, 120 GiB / 120 GiB avail 2026-03-10T10:23:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:29 vm08 ceph-mon[50506]: from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-10T10:23:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:29 vm08 ceph-mon[50506]: osdmap e35: 7 total, 6 up, 7 in 2026-03-10T10:23:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:29 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:23:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:29 vm08 ceph-mon[50506]: from='osd.6 [v2:192.168.123.108:6816/3667011305,v1:192.168.123.108:6817/3667011305]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-10T10:23:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:29 vm08 ceph-mon[50506]: from='osd.6 ' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-10T10:23:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:29 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:23:29.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:29 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:23:29.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:29 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:29.845 INFO:journalctl@ceph.osd.6.vm08.stdout:Mar 10 10:23:29 vm08 ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f-osd-6[60100]: 2026-03-10T10:23:29.477+0000 7fd6a963b700 -1 osd.6 0 waiting for initial osdmap 2026-03-10T10:23:29.845 INFO:journalctl@ceph.osd.6.vm08.stdout:Mar 10 10:23:29 vm08 ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f-osd-6[60100]: 2026-03-10T10:23:29.486+0000 7fd6a37d1700 -1 osd.6 36 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T10:23:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:29 vm01 ceph-mon[48597]: pgmap v61: 1 pgs: 1 active+recovering; 449 KiB data, 35 MiB used, 120 GiB / 120 GiB avail 2026-03-10T10:23:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:29 vm01 ceph-mon[48597]: from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-10T10:23:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:29 vm01 ceph-mon[48597]: osdmap e35: 7 total, 6 up, 7 in 2026-03-10T10:23:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:29 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:23:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:29 vm01 ceph-mon[48597]: from='osd.6 [v2:192.168.123.108:6816/3667011305,v1:192.168.123.108:6817/3667011305]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-10T10:23:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:29 vm01 ceph-mon[48597]: from='osd.6 ' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-10T10:23:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:29 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:23:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:29 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:23:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:29 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:30.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:30 vm08 ceph-mon[50506]: from='client.24205 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm08:/dev/vdb", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:23:30.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:30 vm08 ceph-mon[50506]: from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-10T10:23:30.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:30 vm08 ceph-mon[50506]: osdmap e36: 7 total, 6 up, 7 in 2026-03-10T10:23:30.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:30 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:23:30.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:30 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:30.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:30 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:23:30.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:30 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:23:30.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:30 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:30.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:30 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:23:30.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:30 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:30.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:30 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:23:30.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:30 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:30.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:30 vm08 ceph-mon[50506]: from='client.? 192.168.123.108:0/1968340196' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "1fa744fb-5451-4a33-8f45-c4baa407e004"}]: dispatch 2026-03-10T10:23:30.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:30 vm08 ceph-mon[50506]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "1fa744fb-5451-4a33-8f45-c4baa407e004"}]: dispatch 2026-03-10T10:23:30.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:30 vm08 ceph-mon[50506]: osd.6 [v2:192.168.123.108:6816/3667011305,v1:192.168.123.108:6817/3667011305] boot 2026-03-10T10:23:30.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:30 vm08 ceph-mon[50506]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "1fa744fb-5451-4a33-8f45-c4baa407e004"}]': finished 2026-03-10T10:23:30.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:30 vm08 ceph-mon[50506]: osdmap e37: 8 total, 7 up, 8 in 2026-03-10T10:23:30.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:30 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:23:30.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:30 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:23:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:30 vm01 ceph-mon[48597]: from='client.24205 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm08:/dev/vdb", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:23:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:30 vm01 ceph-mon[48597]: from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-10T10:23:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:30 vm01 ceph-mon[48597]: osdmap e36: 7 total, 6 up, 7 in 2026-03-10T10:23:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:30 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:23:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:30 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:30 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:23:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:30 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:23:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:30 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:30 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:23:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:30 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:30 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:23:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:30 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:30 vm01 ceph-mon[48597]: from='client.? 192.168.123.108:0/1968340196' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "1fa744fb-5451-4a33-8f45-c4baa407e004"}]: dispatch 2026-03-10T10:23:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:30 vm01 ceph-mon[48597]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "1fa744fb-5451-4a33-8f45-c4baa407e004"}]: dispatch 2026-03-10T10:23:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:30 vm01 ceph-mon[48597]: osd.6 [v2:192.168.123.108:6816/3667011305,v1:192.168.123.108:6817/3667011305] boot 2026-03-10T10:23:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:30 vm01 ceph-mon[48597]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "1fa744fb-5451-4a33-8f45-c4baa407e004"}]': finished 2026-03-10T10:23:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:30 vm01 ceph-mon[48597]: osdmap e37: 8 total, 7 up, 8 in 2026-03-10T10:23:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:30 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:23:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:30 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:23:31.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:31 vm08 ceph-mon[50506]: purged_snaps scrub starts 2026-03-10T10:23:31.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:31 vm08 ceph-mon[50506]: purged_snaps scrub ok 2026-03-10T10:23:31.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:31 vm08 ceph-mon[50506]: Detected new or changed devices on vm08 2026-03-10T10:23:31.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:31 vm08 ceph-mon[50506]: pgmap v65: 1 pgs: 1 active+recovering; 449 KiB data, 35 MiB used, 120 GiB / 120 GiB avail 2026-03-10T10:23:31.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:31 vm08 ceph-mon[50506]: from='client.? 192.168.123.108:0/3599974997' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:23:31.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:31 vm08 ceph-mon[50506]: osdmap e38: 8 total, 7 up, 8 in 2026-03-10T10:23:31.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:31 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:23:31.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:31 vm01 ceph-mon[48597]: purged_snaps scrub starts 2026-03-10T10:23:31.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:31 vm01 ceph-mon[48597]: purged_snaps scrub ok 2026-03-10T10:23:31.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:31 vm01 ceph-mon[48597]: Detected new or changed devices on vm08 2026-03-10T10:23:31.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:31 vm01 ceph-mon[48597]: pgmap v65: 1 pgs: 1 active+recovering; 449 KiB data, 35 MiB used, 120 GiB / 120 GiB avail 2026-03-10T10:23:31.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:31 vm01 ceph-mon[48597]: from='client.? 192.168.123.108:0/3599974997' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:23:31.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:31 vm01 ceph-mon[48597]: osdmap e38: 8 total, 7 up, 8 in 2026-03-10T10:23:31.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:31 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:23:33.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:33 vm08 ceph-mon[50506]: osdmap e39: 8 total, 7 up, 8 in 2026-03-10T10:23:33.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:33 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:23:33.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:33 vm08 ceph-mon[50506]: pgmap v68: 1 pgs: 1 remapped+peering; 449 KiB data, 40 MiB used, 140 GiB / 140 GiB avail 2026-03-10T10:23:33.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:33 vm01 ceph-mon[48597]: osdmap e39: 8 total, 7 up, 8 in 2026-03-10T10:23:33.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:33 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:23:33.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:33 vm01 ceph-mon[48597]: pgmap v68: 1 pgs: 1 remapped+peering; 449 KiB data, 40 MiB used, 140 GiB / 140 GiB avail 2026-03-10T10:23:34.333 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:34 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-10T10:23:34.333 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:34 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:34.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:34 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-10T10:23:34.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:34 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:35.209 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:35 vm08 ceph-mon[50506]: Deploying daemon osd.7 on vm08 2026-03-10T10:23:35.209 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:35 vm08 ceph-mon[50506]: pgmap v69: 1 pgs: 1 remapped+peering; 449 KiB data, 41 MiB used, 140 GiB / 140 GiB avail 2026-03-10T10:23:35.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:35 vm01 ceph-mon[48597]: Deploying daemon osd.7 on vm08 2026-03-10T10:23:35.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:35 vm01 ceph-mon[48597]: pgmap v69: 1 pgs: 1 remapped+peering; 449 KiB data, 41 MiB used, 140 GiB / 140 GiB avail 2026-03-10T10:23:36.404 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:36 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:36.405 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:36 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:23:36.405 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:36 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:36.405 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:36 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:23:36.423 INFO:teuthology.orchestra.run.vm08.stdout:Created osd(s) 7 on host 'vm08' 2026-03-10T10:23:36.533 DEBUG:teuthology.orchestra.run.vm08:osd.7> sudo journalctl -f -n 0 -u ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f@osd.7.service 2026-03-10T10:23:36.537 INFO:tasks.cephadm:Waiting for 8 OSDs to come up... 2026-03-10T10:23:36.537 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph osd stat -f json 2026-03-10T10:23:36.558 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:36 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:36.558 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:36 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:23:36.558 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:36 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:36.558 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:36 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:23:36.998 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T10:23:37.059 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":39,"num_osds":8,"num_up_osds":7,"osd_up_since":1773138210,"num_in_osds":8,"osd_in_since":1773138210,"num_remapped_pgs":0} 2026-03-10T10:23:37.094 INFO:journalctl@ceph.osd.7.vm08.stdout:Mar 10 10:23:36 vm08 ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f-osd-7[62915]: 2026-03-10T10:23:36.865+0000 7fd3cd7da3c0 -1 osd.7 0 log_to_monitors true 2026-03-10T10:23:37.673 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:37 vm08 ceph-mon[50506]: pgmap v70: 1 pgs: 1 active+recovering; 449 KiB data, 41 MiB used, 140 GiB / 140 GiB avail; 1/6 objects misplaced (16.667%) 2026-03-10T10:23:37.673 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:37 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:37.673 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:37 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:37.673 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:37 vm08 ceph-mon[50506]: from='osd.7 [v2:192.168.123.108:6824/1668920876,v1:192.168.123.108:6825/1668920876]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-10T10:23:37.673 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:37 vm08 ceph-mon[50506]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-10T10:23:37.673 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:37 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/379326350' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:23:37.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:37 vm01 ceph-mon[48597]: pgmap v70: 1 pgs: 1 active+recovering; 449 KiB data, 41 MiB used, 140 GiB / 140 GiB avail; 1/6 objects misplaced (16.667%) 2026-03-10T10:23:37.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:37 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:37.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:37 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:37.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:37 vm01 ceph-mon[48597]: from='osd.7 [v2:192.168.123.108:6824/1668920876,v1:192.168.123.108:6825/1668920876]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-10T10:23:37.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:37 vm01 ceph-mon[48597]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-10T10:23:37.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:37 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/379326350' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:23:38.060 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph osd stat -f json 2026-03-10T10:23:38.523 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T10:23:38.583 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":40,"num_osds":8,"num_up_osds":7,"osd_up_since":1773138210,"num_in_osds":8,"osd_in_since":1773138210,"num_remapped_pgs":0} 2026-03-10T10:23:38.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:38 vm01 ceph-mon[48597]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-10T10:23:38.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:38 vm01 ceph-mon[48597]: osdmap e40: 8 total, 7 up, 8 in 2026-03-10T10:23:38.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:23:38.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:38 vm01 ceph-mon[48597]: from='osd.7 [v2:192.168.123.108:6824/1668920876,v1:192.168.123.108:6825/1668920876]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-10T10:23:38.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:38 vm01 ceph-mon[48597]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-10T10:23:38.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:38 vm01 ceph-mon[48597]: Detected new or changed devices on vm08 2026-03-10T10:23:38.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:38.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:23:38.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:38.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:23:38.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:38.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:23:38.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:38 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:38.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:38 vm01 ceph-mon[48597]: pgmap v72: 1 pgs: 1 active+recovering; 449 KiB data, 41 MiB used, 140 GiB / 140 GiB avail; 1/6 objects misplaced (16.667%); 9.8 KiB/s, 0 objects/s recovering 2026-03-10T10:23:38.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:38 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/3907239274' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:23:38.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:38 vm08 ceph-mon[50506]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-10T10:23:38.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:38 vm08 ceph-mon[50506]: osdmap e40: 8 total, 7 up, 8 in 2026-03-10T10:23:38.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:23:38.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:38 vm08 ceph-mon[50506]: from='osd.7 [v2:192.168.123.108:6824/1668920876,v1:192.168.123.108:6825/1668920876]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-10T10:23:38.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:38 vm08 ceph-mon[50506]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-10T10:23:38.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:38 vm08 ceph-mon[50506]: Detected new or changed devices on vm08 2026-03-10T10:23:38.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:38.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:23:38.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:38.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:23:38.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:38.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:23:38.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:38 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:38.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:38 vm08 ceph-mon[50506]: pgmap v72: 1 pgs: 1 active+recovering; 449 KiB data, 41 MiB used, 140 GiB / 140 GiB avail; 1/6 objects misplaced (16.667%); 9.8 KiB/s, 0 objects/s recovering 2026-03-10T10:23:38.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:38 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/3907239274' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:23:38.844 INFO:journalctl@ceph.osd.7.vm08.stdout:Mar 10 10:23:38 vm08 ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f-osd-7[62915]: 2026-03-10T10:23:38.549+0000 7fd3c59e0700 -1 osd.7 0 waiting for initial osdmap 2026-03-10T10:23:38.844 INFO:journalctl@ceph.osd.7.vm08.stdout:Mar 10 10:23:38 vm08 ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f-osd-7[62915]: 2026-03-10T10:23:38.568+0000 7fd3beb74700 -1 osd.7 41 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T10:23:39.583 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph osd stat -f json 2026-03-10T10:23:39.821 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:39 vm01 ceph-mon[48597]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-10T10:23:39.821 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:39 vm01 ceph-mon[48597]: osdmap e41: 8 total, 7 up, 8 in 2026-03-10T10:23:39.821 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:39 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:23:39.821 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:39 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:23:39.821 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:39 vm01 ceph-mon[48597]: osd.7 [v2:192.168.123.108:6824/1668920876,v1:192.168.123.108:6825/1668920876] boot 2026-03-10T10:23:39.821 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:39 vm01 ceph-mon[48597]: osdmap e42: 8 total, 8 up, 8 in 2026-03-10T10:23:39.821 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:39 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:23:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:39 vm08 ceph-mon[50506]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-10T10:23:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:39 vm08 ceph-mon[50506]: osdmap e41: 8 total, 7 up, 8 in 2026-03-10T10:23:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:39 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:23:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:39 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:23:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:39 vm08 ceph-mon[50506]: osd.7 [v2:192.168.123.108:6824/1668920876,v1:192.168.123.108:6825/1668920876] boot 2026-03-10T10:23:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:39 vm08 ceph-mon[50506]: osdmap e42: 8 total, 8 up, 8 in 2026-03-10T10:23:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:39 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:23:40.008 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T10:23:40.071 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":42,"num_osds":8,"num_up_osds":8,"osd_up_since":1773138219,"num_in_osds":8,"osd_in_since":1773138210,"num_remapped_pgs":1} 2026-03-10T10:23:40.071 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph osd dump --format=json 2026-03-10T10:23:40.206 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a6b3e19a-1c6a-11f1-aaba-ff800f171c2f/mon.vm01/config 2026-03-10T10:23:40.511 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T10:23:40.512 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":43,"fsid":"a6b3e19a-1c6a-11f1-aaba-ff800f171c2f","created":"2026-03-10T10:21:12.663206+0000","modified":"2026-03-10T10:23:40.037060+0000","last_up_change":"2026-03-10T10:23:39.034671+0000","last_in_change":"2026-03-10T10:23:30.023242+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":18,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"quincy","pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-10T10:22:58.237065+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"20","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}}}],"osds":[{"osd":0,"uuid":"ebe7975c-d06c-485f-8d2b-ebf217684065","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":9,"up_thru":42,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6802","nonce":3587270067},{"type":"v1","addr":"192.168.123.101:6803","nonce":3587270067}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6804","nonce":3587270067},{"type":"v1","addr":"192.168.123.101:6805","nonce":3587270067}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6808","nonce":3587270067},{"type":"v1","addr":"192.168.123.101:6809","nonce":3587270067}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6806","nonce":3587270067},{"type":"v1","addr":"192.168.123.101:6807","nonce":3587270067}]},"public_addr":"192.168.123.101:6803/3587270067","cluster_addr":"192.168.123.101:6805/3587270067","heartbeat_back_addr":"192.168.123.101:6809/3587270067","heartbeat_front_addr":"192.168.123.101:6807/3587270067","state":["exists","up"]},{"osd":1,"uuid":"5fd2081b-4e93-4c56-8dde-73e86252e942","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":28,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6810","nonce":3090715425},{"type":"v1","addr":"192.168.123.101:6811","nonce":3090715425}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6812","nonce":3090715425},{"type":"v1","addr":"192.168.123.101:6813","nonce":3090715425}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6816","nonce":3090715425},{"type":"v1","addr":"192.168.123.101:6817","nonce":3090715425}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6814","nonce":3090715425},{"type":"v1","addr":"192.168.123.101:6815","nonce":3090715425}]},"public_addr":"192.168.123.101:6811/3090715425","cluster_addr":"192.168.123.101:6813/3090715425","heartbeat_back_addr":"192.168.123.101:6817/3090715425","heartbeat_front_addr":"192.168.123.101:6815/3090715425","state":["exists","up"]},{"osd":2,"uuid":"7b14edda-e6e2-4ff6-aa30-d875c9b27b49","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6818","nonce":1447659236},{"type":"v1","addr":"192.168.123.101:6819","nonce":1447659236}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6820","nonce":1447659236},{"type":"v1","addr":"192.168.123.101:6821","nonce":1447659236}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6824","nonce":1447659236},{"type":"v1","addr":"192.168.123.101:6825","nonce":1447659236}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6822","nonce":1447659236},{"type":"v1","addr":"192.168.123.101:6823","nonce":1447659236}]},"public_addr":"192.168.123.101:6819/1447659236","cluster_addr":"192.168.123.101:6821/1447659236","heartbeat_back_addr":"192.168.123.101:6825/1447659236","heartbeat_front_addr":"192.168.123.101:6823/1447659236","state":["exists","up"]},{"osd":3,"uuid":"af0a15a7-05e5-4909-8944-efe4cc34a49e","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":23,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6826","nonce":3038741761},{"type":"v1","addr":"192.168.123.101:6827","nonce":3038741761}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6828","nonce":3038741761},{"type":"v1","addr":"192.168.123.101:6829","nonce":3038741761}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6832","nonce":3038741761},{"type":"v1","addr":"192.168.123.101:6833","nonce":3038741761}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6830","nonce":3038741761},{"type":"v1","addr":"192.168.123.101:6831","nonce":3038741761}]},"public_addr":"192.168.123.101:6827/3038741761","cluster_addr":"192.168.123.101:6829/3038741761","heartbeat_back_addr":"192.168.123.101:6833/3038741761","heartbeat_front_addr":"192.168.123.101:6831/3038741761","state":["exists","up"]},{"osd":4,"uuid":"b7907d54-5467-4a9f-ae4f-614fa10c8000","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":27,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6800","nonce":3235322654},{"type":"v1","addr":"192.168.123.108:6801","nonce":3235322654}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6802","nonce":3235322654},{"type":"v1","addr":"192.168.123.108:6803","nonce":3235322654}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6806","nonce":3235322654},{"type":"v1","addr":"192.168.123.108:6807","nonce":3235322654}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6804","nonce":3235322654},{"type":"v1","addr":"192.168.123.108:6805","nonce":3235322654}]},"public_addr":"192.168.123.108:6801/3235322654","cluster_addr":"192.168.123.108:6803/3235322654","heartbeat_back_addr":"192.168.123.108:6807/3235322654","heartbeat_front_addr":"192.168.123.108:6805/3235322654","state":["exists","up"]},{"osd":5,"uuid":"3a21cd68-d40c-4c3b-a20b-efb92c0277a1","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":32,"up_thru":33,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6808","nonce":2436937416},{"type":"v1","addr":"192.168.123.108:6809","nonce":2436937416}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6810","nonce":2436937416},{"type":"v1","addr":"192.168.123.108:6811","nonce":2436937416}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6814","nonce":2436937416},{"type":"v1","addr":"192.168.123.108:6815","nonce":2436937416}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6812","nonce":2436937416},{"type":"v1","addr":"192.168.123.108:6813","nonce":2436937416}]},"public_addr":"192.168.123.108:6809/2436937416","cluster_addr":"192.168.123.108:6811/2436937416","heartbeat_back_addr":"192.168.123.108:6815/2436937416","heartbeat_front_addr":"192.168.123.108:6813/2436937416","state":["exists","up"]},{"osd":6,"uuid":"162defaf-0c11-4032-b25b-1848a829ed6f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":37,"up_thru":38,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6816","nonce":3667011305},{"type":"v1","addr":"192.168.123.108:6817","nonce":3667011305}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6818","nonce":3667011305},{"type":"v1","addr":"192.168.123.108:6819","nonce":3667011305}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6822","nonce":3667011305},{"type":"v1","addr":"192.168.123.108:6823","nonce":3667011305}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6820","nonce":3667011305},{"type":"v1","addr":"192.168.123.108:6821","nonce":3667011305}]},"public_addr":"192.168.123.108:6817/3667011305","cluster_addr":"192.168.123.108:6819/3667011305","heartbeat_back_addr":"192.168.123.108:6823/3667011305","heartbeat_front_addr":"192.168.123.108:6821/3667011305","state":["exists","up"]},{"osd":7,"uuid":"1fa744fb-5451-4a33-8f45-c4baa407e004","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":42,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6824","nonce":1668920876},{"type":"v1","addr":"192.168.123.108:6825","nonce":1668920876}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6826","nonce":1668920876},{"type":"v1","addr":"192.168.123.108:6827","nonce":1668920876}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6830","nonce":1668920876},{"type":"v1","addr":"192.168.123.108:6831","nonce":1668920876}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6828","nonce":1668920876},{"type":"v1","addr":"192.168.123.108:6829","nonce":1668920876}]},"public_addr":"192.168.123.108:6825/1668920876","cluster_addr":"192.168.123.108:6827/1668920876","heartbeat_back_addr":"192.168.123.108:6831/1668920876","heartbeat_front_addr":"192.168.123.108:6829/1668920876","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:22:38.863960+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:22:47.329498+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:22:56.683013+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:23:05.436916+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:23:13.071310+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:23:21.132598+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:23:28.856576+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:23:37.818488+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.101:0/1055495161":"2026-03-11T10:22:09.576438+0000","192.168.123.101:6800/2333302559":"2026-03-11T10:22:09.576438+0000","192.168.123.101:0/3872135780":"2026-03-11T10:21:39.818421+0000","192.168.123.101:0/2605615659":"2026-03-11T10:21:39.818421+0000","192.168.123.101:6801/2333302559":"2026-03-11T10:22:09.576438+0000","192.168.123.101:0/2758406695":"2026-03-11T10:21:27.174153+0000","192.168.123.101:6801/2830381421":"2026-03-11T10:21:39.818421+0000","192.168.123.101:0/1747073541":"2026-03-11T10:21:39.818421+0000","192.168.123.101:6800/660144235":"2026-03-11T10:21:27.174153+0000","192.168.123.101:0/2531604582":"2026-03-11T10:22:09.576438+0000","192.168.123.101:6801/660144235":"2026-03-11T10:21:27.174153+0000","192.168.123.101:6800/2830381421":"2026-03-11T10:21:39.818421+0000","192.168.123.101:0/930880465":"2026-03-11T10:21:27.174153+0000","192.168.123.101:0/1010726127":"2026-03-11T10:22:09.576438+0000","192.168.123.101:0/460595964":"2026-03-11T10:21:27.174153+0000"},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-10T10:23:40.614 INFO:tasks.cephadm.ceph_manager.ceph:[{'pool': 1, 'pool_name': '.mgr', 'create_time': '2026-03-10T10:22:58.237065+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'off', 'pg_num': 1, 'pg_placement_num': 1, 'pg_placement_num_target': 1, 'pg_num_target': 1, 'pg_num_pending': 1, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': "0'0", 'target_version': "0'0"}, 'last_change': '20', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {'pg_num_max': 32, 'pg_num_min': 1}, 'application_metadata': {'mgr': {}}}] 2026-03-10T10:23:40.614 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph osd pool get .mgr pg_num 2026-03-10T10:23:40.762 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a6b3e19a-1c6a-11f1-aaba-ff800f171c2f/mon.vm01/config 2026-03-10T10:23:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:40 vm08 ceph-mon[50506]: purged_snaps scrub starts 2026-03-10T10:23:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:40 vm08 ceph-mon[50506]: purged_snaps scrub ok 2026-03-10T10:23:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:40 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/3709577064' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:23:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:40 vm08 ceph-mon[50506]: osdmap e43: 8 total, 8 up, 8 in 2026-03-10T10:23:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:40 vm08 ceph-mon[50506]: pgmap v76: 1 pgs: 1 active+recovering; 449 KiB data, 46 MiB used, 160 GiB / 160 GiB avail; 1/6 objects misplaced (16.667%) 2026-03-10T10:23:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:40 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/1056444655' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:23:40.851 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:40 vm01 ceph-mon[48597]: purged_snaps scrub starts 2026-03-10T10:23:40.851 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:40 vm01 ceph-mon[48597]: purged_snaps scrub ok 2026-03-10T10:23:40.851 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:40 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/3709577064' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:23:40.851 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:40 vm01 ceph-mon[48597]: osdmap e43: 8 total, 8 up, 8 in 2026-03-10T10:23:40.851 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:40 vm01 ceph-mon[48597]: pgmap v76: 1 pgs: 1 active+recovering; 449 KiB data, 46 MiB used, 160 GiB / 160 GiB avail; 1/6 objects misplaced (16.667%) 2026-03-10T10:23:40.851 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:40 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/1056444655' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:23:41.089 INFO:teuthology.orchestra.run.vm01.stdout:pg_num: 1 2026-03-10T10:23:41.146 INFO:tasks.cephadm:Setting up client nodes... 2026-03-10T10:23:41.146 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph auth get-or-create client.0 mon 'allow *' osd 'allow *' mds 'allow *' mgr 'allow *' 2026-03-10T10:23:41.578 INFO:teuthology.orchestra.run.vm01.stdout:[client.0] 2026-03-10T10:23:41.578 INFO:teuthology.orchestra.run.vm01.stdout: key = AQAt8a9pDqgwIhAAa2o8322lkVP2BrlGbHcSLw== 2026-03-10T10:23:41.639 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T10:23:41.639 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/ceph/ceph.client.0.keyring 2026-03-10T10:23:41.639 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod 0644 /etc/ceph/ceph.client.0.keyring 2026-03-10T10:23:41.670 INFO:tasks.ceph:Waiting until ceph daemons up and pgs clean... 2026-03-10T10:23:41.670 INFO:tasks.cephadm.ceph_manager.ceph:waiting for mgr available 2026-03-10T10:23:41.670 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph mgr dump --format=json 2026-03-10T10:23:41.843 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a6b3e19a-1c6a-11f1-aaba-ff800f171c2f/mon.vm01/config 2026-03-10T10:23:42.154 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:42 vm01 ceph-mon[48597]: osdmap e44: 8 total, 8 up, 8 in 2026-03-10T10:23:42.154 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:42 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/3617437087' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-10T10:23:42.154 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:42 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/3010156055' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T10:23:42.154 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:42 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/3010156055' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-10T10:23:42.154 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T10:23:42.218 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":19,"active_gid":14216,"active_name":"vm01.nkkzmc","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6800","nonce":3098475737},{"type":"v1","addr":"192.168.123.101:6801","nonce":3098475737}]},"active_addr":"192.168.123.101:6801/3098475737","active_change":"2026-03-10T10:22:09.606753+0000","active_mgr_features":4540138303579357183,"available":true,"standbys":[{"gid":14234,"name":"vm08.rwpjaf","mgr_features":4540138303579357183,"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"7","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 or 7 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2400","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"7","min":"0","max":"7","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 or 7 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","upmap"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.23.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/ceph-grafana:8.3.5","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"docker.io/library/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"docker.io/arcts/keepalived","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.3.1","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.33.4","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"docker.io/maxwo/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"docker.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"noautoscale":{"name":"noautoscale","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"global autoscale flag","long_desc":"Option to turn on/off the autoscaler for all pools","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"drive_group_interval":{"name":"drive_group_interval","type":"float","level":"advanced","flags":0,"default_value":"300.0","min":"","max":"","enum_allowed":[],"desc":"interval in seconds between re-application of applied drive_groups","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}]}],"modules":["cephadm","dashboard","iostat","nfs","prometheus","restful"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"7","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 or 7 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2400","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"7","min":"0","max":"7","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 or 7 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","upmap"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.23.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/ceph-grafana:8.3.5","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"docker.io/library/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"docker.io/arcts/keepalived","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.3.1","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.33.4","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"docker.io/maxwo/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"docker.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"noautoscale":{"name":"noautoscale","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"global autoscale flag","long_desc":"Option to turn on/off the autoscaler for all pools","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"drive_group_interval":{"name":"drive_group_interval","type":"float","level":"advanced","flags":0,"default_value":"300.0","min":"","max":"","enum_allowed":[],"desc":"interval in seconds between re-application of applied drive_groups","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{"dashboard":"https://192.168.123.101:8443/","prometheus":"http://192.168.123.101:9283/"},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"last_failure_osd_epoch":5,"active_clients":[{"addrvec":[{"type":"v2","addr":"192.168.123.101:0","nonce":3061044045}]},{"addrvec":[{"type":"v2","addr":"192.168.123.101:0","nonce":2170596599}]},{"addrvec":[{"type":"v2","addr":"192.168.123.101:0","nonce":3768853481}]},{"addrvec":[{"type":"v2","addr":"192.168.123.101:0","nonce":4213450286}]}]}} 2026-03-10T10:23:42.219 INFO:tasks.cephadm.ceph_manager.ceph:mgr available! 2026-03-10T10:23:42.219 INFO:tasks.cephadm.ceph_manager.ceph:waiting for all up 2026-03-10T10:23:42.219 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph osd dump --format=json 2026-03-10T10:23:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:42 vm08 ceph-mon[50506]: osdmap e44: 8 total, 8 up, 8 in 2026-03-10T10:23:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:42 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/3617437087' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-10T10:23:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:42 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/3010156055' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T10:23:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:42 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/3010156055' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-10T10:23:42.351 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a6b3e19a-1c6a-11f1-aaba-ff800f171c2f/mon.vm01/config 2026-03-10T10:23:42.641 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T10:23:42.641 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":44,"fsid":"a6b3e19a-1c6a-11f1-aaba-ff800f171c2f","created":"2026-03-10T10:21:12.663206+0000","modified":"2026-03-10T10:23:41.040454+0000","last_up_change":"2026-03-10T10:23:39.034671+0000","last_in_change":"2026-03-10T10:23:30.023242+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":18,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"quincy","pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-10T10:22:58.237065+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"20","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}}}],"osds":[{"osd":0,"uuid":"ebe7975c-d06c-485f-8d2b-ebf217684065","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":9,"up_thru":42,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6802","nonce":3587270067},{"type":"v1","addr":"192.168.123.101:6803","nonce":3587270067}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6804","nonce":3587270067},{"type":"v1","addr":"192.168.123.101:6805","nonce":3587270067}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6808","nonce":3587270067},{"type":"v1","addr":"192.168.123.101:6809","nonce":3587270067}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6806","nonce":3587270067},{"type":"v1","addr":"192.168.123.101:6807","nonce":3587270067}]},"public_addr":"192.168.123.101:6803/3587270067","cluster_addr":"192.168.123.101:6805/3587270067","heartbeat_back_addr":"192.168.123.101:6809/3587270067","heartbeat_front_addr":"192.168.123.101:6807/3587270067","state":["exists","up"]},{"osd":1,"uuid":"5fd2081b-4e93-4c56-8dde-73e86252e942","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":28,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6810","nonce":3090715425},{"type":"v1","addr":"192.168.123.101:6811","nonce":3090715425}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6812","nonce":3090715425},{"type":"v1","addr":"192.168.123.101:6813","nonce":3090715425}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6816","nonce":3090715425},{"type":"v1","addr":"192.168.123.101:6817","nonce":3090715425}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6814","nonce":3090715425},{"type":"v1","addr":"192.168.123.101:6815","nonce":3090715425}]},"public_addr":"192.168.123.101:6811/3090715425","cluster_addr":"192.168.123.101:6813/3090715425","heartbeat_back_addr":"192.168.123.101:6817/3090715425","heartbeat_front_addr":"192.168.123.101:6815/3090715425","state":["exists","up"]},{"osd":2,"uuid":"7b14edda-e6e2-4ff6-aa30-d875c9b27b49","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6818","nonce":1447659236},{"type":"v1","addr":"192.168.123.101:6819","nonce":1447659236}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6820","nonce":1447659236},{"type":"v1","addr":"192.168.123.101:6821","nonce":1447659236}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6824","nonce":1447659236},{"type":"v1","addr":"192.168.123.101:6825","nonce":1447659236}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6822","nonce":1447659236},{"type":"v1","addr":"192.168.123.101:6823","nonce":1447659236}]},"public_addr":"192.168.123.101:6819/1447659236","cluster_addr":"192.168.123.101:6821/1447659236","heartbeat_back_addr":"192.168.123.101:6825/1447659236","heartbeat_front_addr":"192.168.123.101:6823/1447659236","state":["exists","up"]},{"osd":3,"uuid":"af0a15a7-05e5-4909-8944-efe4cc34a49e","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":23,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6826","nonce":3038741761},{"type":"v1","addr":"192.168.123.101:6827","nonce":3038741761}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6828","nonce":3038741761},{"type":"v1","addr":"192.168.123.101:6829","nonce":3038741761}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6832","nonce":3038741761},{"type":"v1","addr":"192.168.123.101:6833","nonce":3038741761}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6830","nonce":3038741761},{"type":"v1","addr":"192.168.123.101:6831","nonce":3038741761}]},"public_addr":"192.168.123.101:6827/3038741761","cluster_addr":"192.168.123.101:6829/3038741761","heartbeat_back_addr":"192.168.123.101:6833/3038741761","heartbeat_front_addr":"192.168.123.101:6831/3038741761","state":["exists","up"]},{"osd":4,"uuid":"b7907d54-5467-4a9f-ae4f-614fa10c8000","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":27,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6800","nonce":3235322654},{"type":"v1","addr":"192.168.123.108:6801","nonce":3235322654}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6802","nonce":3235322654},{"type":"v1","addr":"192.168.123.108:6803","nonce":3235322654}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6806","nonce":3235322654},{"type":"v1","addr":"192.168.123.108:6807","nonce":3235322654}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6804","nonce":3235322654},{"type":"v1","addr":"192.168.123.108:6805","nonce":3235322654}]},"public_addr":"192.168.123.108:6801/3235322654","cluster_addr":"192.168.123.108:6803/3235322654","heartbeat_back_addr":"192.168.123.108:6807/3235322654","heartbeat_front_addr":"192.168.123.108:6805/3235322654","state":["exists","up"]},{"osd":5,"uuid":"3a21cd68-d40c-4c3b-a20b-efb92c0277a1","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":32,"up_thru":33,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6808","nonce":2436937416},{"type":"v1","addr":"192.168.123.108:6809","nonce":2436937416}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6810","nonce":2436937416},{"type":"v1","addr":"192.168.123.108:6811","nonce":2436937416}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6814","nonce":2436937416},{"type":"v1","addr":"192.168.123.108:6815","nonce":2436937416}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6812","nonce":2436937416},{"type":"v1","addr":"192.168.123.108:6813","nonce":2436937416}]},"public_addr":"192.168.123.108:6809/2436937416","cluster_addr":"192.168.123.108:6811/2436937416","heartbeat_back_addr":"192.168.123.108:6815/2436937416","heartbeat_front_addr":"192.168.123.108:6813/2436937416","state":["exists","up"]},{"osd":6,"uuid":"162defaf-0c11-4032-b25b-1848a829ed6f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":37,"up_thru":38,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6816","nonce":3667011305},{"type":"v1","addr":"192.168.123.108:6817","nonce":3667011305}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6818","nonce":3667011305},{"type":"v1","addr":"192.168.123.108:6819","nonce":3667011305}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6822","nonce":3667011305},{"type":"v1","addr":"192.168.123.108:6823","nonce":3667011305}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6820","nonce":3667011305},{"type":"v1","addr":"192.168.123.108:6821","nonce":3667011305}]},"public_addr":"192.168.123.108:6817/3667011305","cluster_addr":"192.168.123.108:6819/3667011305","heartbeat_back_addr":"192.168.123.108:6823/3667011305","heartbeat_front_addr":"192.168.123.108:6821/3667011305","state":["exists","up"]},{"osd":7,"uuid":"1fa744fb-5451-4a33-8f45-c4baa407e004","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":42,"up_thru":43,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6824","nonce":1668920876},{"type":"v1","addr":"192.168.123.108:6825","nonce":1668920876}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6826","nonce":1668920876},{"type":"v1","addr":"192.168.123.108:6827","nonce":1668920876}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6830","nonce":1668920876},{"type":"v1","addr":"192.168.123.108:6831","nonce":1668920876}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6828","nonce":1668920876},{"type":"v1","addr":"192.168.123.108:6829","nonce":1668920876}]},"public_addr":"192.168.123.108:6825/1668920876","cluster_addr":"192.168.123.108:6827/1668920876","heartbeat_back_addr":"192.168.123.108:6831/1668920876","heartbeat_front_addr":"192.168.123.108:6829/1668920876","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:22:38.863960+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:22:47.329498+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:22:56.683013+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:23:05.436916+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:23:13.071310+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:23:21.132598+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:23:28.856576+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:23:37.818488+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.101:0/1055495161":"2026-03-11T10:22:09.576438+0000","192.168.123.101:6800/2333302559":"2026-03-11T10:22:09.576438+0000","192.168.123.101:0/3872135780":"2026-03-11T10:21:39.818421+0000","192.168.123.101:0/2605615659":"2026-03-11T10:21:39.818421+0000","192.168.123.101:6801/2333302559":"2026-03-11T10:22:09.576438+0000","192.168.123.101:0/2758406695":"2026-03-11T10:21:27.174153+0000","192.168.123.101:6801/2830381421":"2026-03-11T10:21:39.818421+0000","192.168.123.101:0/1747073541":"2026-03-11T10:21:39.818421+0000","192.168.123.101:6800/660144235":"2026-03-11T10:21:27.174153+0000","192.168.123.101:0/2531604582":"2026-03-11T10:22:09.576438+0000","192.168.123.101:6801/660144235":"2026-03-11T10:21:27.174153+0000","192.168.123.101:6800/2830381421":"2026-03-11T10:21:39.818421+0000","192.168.123.101:0/930880465":"2026-03-11T10:21:27.174153+0000","192.168.123.101:0/1010726127":"2026-03-11T10:22:09.576438+0000","192.168.123.101:0/460595964":"2026-03-11T10:21:27.174153+0000"},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-10T10:23:42.702 INFO:tasks.cephadm.ceph_manager.ceph:all up! 2026-03-10T10:23:42.702 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph osd dump --format=json 2026-03-10T10:23:42.834 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a6b3e19a-1c6a-11f1-aaba-ff800f171c2f/mon.vm01/config 2026-03-10T10:23:43.143 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T10:23:43.143 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":44,"fsid":"a6b3e19a-1c6a-11f1-aaba-ff800f171c2f","created":"2026-03-10T10:21:12.663206+0000","modified":"2026-03-10T10:23:41.040454+0000","last_up_change":"2026-03-10T10:23:39.034671+0000","last_in_change":"2026-03-10T10:23:30.023242+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":18,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"quincy","pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-10T10:22:58.237065+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"20","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}}}],"osds":[{"osd":0,"uuid":"ebe7975c-d06c-485f-8d2b-ebf217684065","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":9,"up_thru":42,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6802","nonce":3587270067},{"type":"v1","addr":"192.168.123.101:6803","nonce":3587270067}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6804","nonce":3587270067},{"type":"v1","addr":"192.168.123.101:6805","nonce":3587270067}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6808","nonce":3587270067},{"type":"v1","addr":"192.168.123.101:6809","nonce":3587270067}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6806","nonce":3587270067},{"type":"v1","addr":"192.168.123.101:6807","nonce":3587270067}]},"public_addr":"192.168.123.101:6803/3587270067","cluster_addr":"192.168.123.101:6805/3587270067","heartbeat_back_addr":"192.168.123.101:6809/3587270067","heartbeat_front_addr":"192.168.123.101:6807/3587270067","state":["exists","up"]},{"osd":1,"uuid":"5fd2081b-4e93-4c56-8dde-73e86252e942","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":28,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6810","nonce":3090715425},{"type":"v1","addr":"192.168.123.101:6811","nonce":3090715425}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6812","nonce":3090715425},{"type":"v1","addr":"192.168.123.101:6813","nonce":3090715425}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6816","nonce":3090715425},{"type":"v1","addr":"192.168.123.101:6817","nonce":3090715425}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6814","nonce":3090715425},{"type":"v1","addr":"192.168.123.101:6815","nonce":3090715425}]},"public_addr":"192.168.123.101:6811/3090715425","cluster_addr":"192.168.123.101:6813/3090715425","heartbeat_back_addr":"192.168.123.101:6817/3090715425","heartbeat_front_addr":"192.168.123.101:6815/3090715425","state":["exists","up"]},{"osd":2,"uuid":"7b14edda-e6e2-4ff6-aa30-d875c9b27b49","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6818","nonce":1447659236},{"type":"v1","addr":"192.168.123.101:6819","nonce":1447659236}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6820","nonce":1447659236},{"type":"v1","addr":"192.168.123.101:6821","nonce":1447659236}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6824","nonce":1447659236},{"type":"v1","addr":"192.168.123.101:6825","nonce":1447659236}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6822","nonce":1447659236},{"type":"v1","addr":"192.168.123.101:6823","nonce":1447659236}]},"public_addr":"192.168.123.101:6819/1447659236","cluster_addr":"192.168.123.101:6821/1447659236","heartbeat_back_addr":"192.168.123.101:6825/1447659236","heartbeat_front_addr":"192.168.123.101:6823/1447659236","state":["exists","up"]},{"osd":3,"uuid":"af0a15a7-05e5-4909-8944-efe4cc34a49e","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":23,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6826","nonce":3038741761},{"type":"v1","addr":"192.168.123.101:6827","nonce":3038741761}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6828","nonce":3038741761},{"type":"v1","addr":"192.168.123.101:6829","nonce":3038741761}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6832","nonce":3038741761},{"type":"v1","addr":"192.168.123.101:6833","nonce":3038741761}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6830","nonce":3038741761},{"type":"v1","addr":"192.168.123.101:6831","nonce":3038741761}]},"public_addr":"192.168.123.101:6827/3038741761","cluster_addr":"192.168.123.101:6829/3038741761","heartbeat_back_addr":"192.168.123.101:6833/3038741761","heartbeat_front_addr":"192.168.123.101:6831/3038741761","state":["exists","up"]},{"osd":4,"uuid":"b7907d54-5467-4a9f-ae4f-614fa10c8000","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":27,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6800","nonce":3235322654},{"type":"v1","addr":"192.168.123.108:6801","nonce":3235322654}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6802","nonce":3235322654},{"type":"v1","addr":"192.168.123.108:6803","nonce":3235322654}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6806","nonce":3235322654},{"type":"v1","addr":"192.168.123.108:6807","nonce":3235322654}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6804","nonce":3235322654},{"type":"v1","addr":"192.168.123.108:6805","nonce":3235322654}]},"public_addr":"192.168.123.108:6801/3235322654","cluster_addr":"192.168.123.108:6803/3235322654","heartbeat_back_addr":"192.168.123.108:6807/3235322654","heartbeat_front_addr":"192.168.123.108:6805/3235322654","state":["exists","up"]},{"osd":5,"uuid":"3a21cd68-d40c-4c3b-a20b-efb92c0277a1","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":32,"up_thru":33,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6808","nonce":2436937416},{"type":"v1","addr":"192.168.123.108:6809","nonce":2436937416}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6810","nonce":2436937416},{"type":"v1","addr":"192.168.123.108:6811","nonce":2436937416}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6814","nonce":2436937416},{"type":"v1","addr":"192.168.123.108:6815","nonce":2436937416}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6812","nonce":2436937416},{"type":"v1","addr":"192.168.123.108:6813","nonce":2436937416}]},"public_addr":"192.168.123.108:6809/2436937416","cluster_addr":"192.168.123.108:6811/2436937416","heartbeat_back_addr":"192.168.123.108:6815/2436937416","heartbeat_front_addr":"192.168.123.108:6813/2436937416","state":["exists","up"]},{"osd":6,"uuid":"162defaf-0c11-4032-b25b-1848a829ed6f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":37,"up_thru":38,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6816","nonce":3667011305},{"type":"v1","addr":"192.168.123.108:6817","nonce":3667011305}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6818","nonce":3667011305},{"type":"v1","addr":"192.168.123.108:6819","nonce":3667011305}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6822","nonce":3667011305},{"type":"v1","addr":"192.168.123.108:6823","nonce":3667011305}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6820","nonce":3667011305},{"type":"v1","addr":"192.168.123.108:6821","nonce":3667011305}]},"public_addr":"192.168.123.108:6817/3667011305","cluster_addr":"192.168.123.108:6819/3667011305","heartbeat_back_addr":"192.168.123.108:6823/3667011305","heartbeat_front_addr":"192.168.123.108:6821/3667011305","state":["exists","up"]},{"osd":7,"uuid":"1fa744fb-5451-4a33-8f45-c4baa407e004","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":42,"up_thru":43,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6824","nonce":1668920876},{"type":"v1","addr":"192.168.123.108:6825","nonce":1668920876}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6826","nonce":1668920876},{"type":"v1","addr":"192.168.123.108:6827","nonce":1668920876}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6830","nonce":1668920876},{"type":"v1","addr":"192.168.123.108:6831","nonce":1668920876}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6828","nonce":1668920876},{"type":"v1","addr":"192.168.123.108:6829","nonce":1668920876}]},"public_addr":"192.168.123.108:6825/1668920876","cluster_addr":"192.168.123.108:6827/1668920876","heartbeat_back_addr":"192.168.123.108:6831/1668920876","heartbeat_front_addr":"192.168.123.108:6829/1668920876","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:22:38.863960+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:22:47.329498+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:22:56.683013+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:23:05.436916+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:23:13.071310+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:23:21.132598+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:23:28.856576+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:23:37.818488+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.101:0/1055495161":"2026-03-11T10:22:09.576438+0000","192.168.123.101:6800/2333302559":"2026-03-11T10:22:09.576438+0000","192.168.123.101:0/3872135780":"2026-03-11T10:21:39.818421+0000","192.168.123.101:0/2605615659":"2026-03-11T10:21:39.818421+0000","192.168.123.101:6801/2333302559":"2026-03-11T10:22:09.576438+0000","192.168.123.101:0/2758406695":"2026-03-11T10:21:27.174153+0000","192.168.123.101:6801/2830381421":"2026-03-11T10:21:39.818421+0000","192.168.123.101:0/1747073541":"2026-03-11T10:21:39.818421+0000","192.168.123.101:6800/660144235":"2026-03-11T10:21:27.174153+0000","192.168.123.101:0/2531604582":"2026-03-11T10:22:09.576438+0000","192.168.123.101:6801/660144235":"2026-03-11T10:21:27.174153+0000","192.168.123.101:6800/2830381421":"2026-03-11T10:21:39.818421+0000","192.168.123.101:0/930880465":"2026-03-11T10:21:27.174153+0000","192.168.123.101:0/1010726127":"2026-03-11T10:22:09.576438+0000","192.168.123.101:0/460595964":"2026-03-11T10:21:27.174153+0000"},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-10T10:23:43.201 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:43 vm01 ceph-mon[48597]: pgmap v78: 1 pgs: 1 active+recovering; 449 KiB data, 47 MiB used, 160 GiB / 160 GiB avail; 1/6 objects misplaced (16.667%) 2026-03-10T10:23:43.201 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:43 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/891055021' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-10T10:23:43.201 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:43 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/3952033461' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:23:43.202 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph tell osd.0 flush_pg_stats 2026-03-10T10:23:43.202 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph tell osd.1 flush_pg_stats 2026-03-10T10:23:43.202 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph tell osd.2 flush_pg_stats 2026-03-10T10:23:43.202 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph tell osd.3 flush_pg_stats 2026-03-10T10:23:43.202 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph tell osd.4 flush_pg_stats 2026-03-10T10:23:43.202 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph tell osd.5 flush_pg_stats 2026-03-10T10:23:43.202 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph tell osd.6 flush_pg_stats 2026-03-10T10:23:43.202 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph tell osd.7 flush_pg_stats 2026-03-10T10:23:43.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:43 vm08 ceph-mon[50506]: pgmap v78: 1 pgs: 1 active+recovering; 449 KiB data, 47 MiB used, 160 GiB / 160 GiB avail; 1/6 objects misplaced (16.667%) 2026-03-10T10:23:43.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:43 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/891055021' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-10T10:23:43.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:43 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/3952033461' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:23:43.701 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a6b3e19a-1c6a-11f1-aaba-ff800f171c2f/mon.vm01/config 2026-03-10T10:23:43.720 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a6b3e19a-1c6a-11f1-aaba-ff800f171c2f/mon.vm01/config 2026-03-10T10:23:43.916 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a6b3e19a-1c6a-11f1-aaba-ff800f171c2f/mon.vm01/config 2026-03-10T10:23:43.943 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a6b3e19a-1c6a-11f1-aaba-ff800f171c2f/mon.vm01/config 2026-03-10T10:23:43.945 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a6b3e19a-1c6a-11f1-aaba-ff800f171c2f/mon.vm01/config 2026-03-10T10:23:43.948 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a6b3e19a-1c6a-11f1-aaba-ff800f171c2f/mon.vm01/config 2026-03-10T10:23:44.012 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a6b3e19a-1c6a-11f1-aaba-ff800f171c2f/mon.vm01/config 2026-03-10T10:23:44.020 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a6b3e19a-1c6a-11f1-aaba-ff800f171c2f/mon.vm01/config 2026-03-10T10:23:44.081 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:44 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/3747131544' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:23:44.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:44 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/3747131544' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:23:45.136 INFO:teuthology.orchestra.run.vm01.stdout:98784247817 2026-03-10T10:23:45.136 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph osd last-stat-seq osd.3 2026-03-10T10:23:45.147 INFO:teuthology.orchestra.run.vm01.stdout:180388626435 2026-03-10T10:23:45.147 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph osd last-stat-seq osd.7 2026-03-10T10:23:45.172 INFO:teuthology.orchestra.run.vm01.stdout:38654705678 2026-03-10T10:23:45.172 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph osd last-stat-seq osd.0 2026-03-10T10:23:45.296 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:45 vm01 ceph-mon[48597]: pgmap v79: 1 pgs: 1 active+recovering; 449 KiB data, 47 MiB used, 160 GiB / 160 GiB avail; 1/6 objects misplaced (16.667%) 2026-03-10T10:23:45.338 INFO:teuthology.orchestra.run.vm01.stdout:55834574861 2026-03-10T10:23:45.338 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph osd last-stat-seq osd.1 2026-03-10T10:23:45.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:45 vm08 ceph-mon[50506]: pgmap v79: 1 pgs: 1 active+recovering; 449 KiB data, 47 MiB used, 160 GiB / 160 GiB avail; 1/6 objects misplaced (16.667%) 2026-03-10T10:23:45.360 INFO:teuthology.orchestra.run.vm01.stdout:73014444043 2026-03-10T10:23:45.361 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph osd last-stat-seq osd.2 2026-03-10T10:23:45.441 INFO:teuthology.orchestra.run.vm01.stdout:115964117000 2026-03-10T10:23:45.441 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph osd last-stat-seq osd.4 2026-03-10T10:23:45.461 INFO:teuthology.orchestra.run.vm01.stdout:158913789956 2026-03-10T10:23:45.461 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph osd last-stat-seq osd.6 2026-03-10T10:23:45.469 INFO:teuthology.orchestra.run.vm01.stdout:137438953478 2026-03-10T10:23:45.469 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph osd last-stat-seq osd.5 2026-03-10T10:23:45.646 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a6b3e19a-1c6a-11f1-aaba-ff800f171c2f/mon.vm01/config 2026-03-10T10:23:45.825 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a6b3e19a-1c6a-11f1-aaba-ff800f171c2f/mon.vm01/config 2026-03-10T10:23:45.897 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a6b3e19a-1c6a-11f1-aaba-ff800f171c2f/mon.vm01/config 2026-03-10T10:23:45.940 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a6b3e19a-1c6a-11f1-aaba-ff800f171c2f/mon.vm01/config 2026-03-10T10:23:46.363 INFO:teuthology.orchestra.run.vm01.stdout:98784247817 2026-03-10T10:23:46.481 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a6b3e19a-1c6a-11f1-aaba-ff800f171c2f/mon.vm01/config 2026-03-10T10:23:46.494 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a6b3e19a-1c6a-11f1-aaba-ff800f171c2f/mon.vm01/config 2026-03-10T10:23:46.515 INFO:tasks.cephadm.ceph_manager.ceph:need seq 98784247817 got 98784247817 for osd.3 2026-03-10T10:23:46.516 DEBUG:teuthology.parallel:result is None 2026-03-10T10:23:46.680 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a6b3e19a-1c6a-11f1-aaba-ff800f171c2f/mon.vm01/config 2026-03-10T10:23:46.716 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a6b3e19a-1c6a-11f1-aaba-ff800f171c2f/mon.vm01/config 2026-03-10T10:23:47.175 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:47 vm01 ceph-mon[48597]: pgmap v80: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 46 KiB/s, 0 objects/s recovering 2026-03-10T10:23:47.175 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:47 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/1791560253' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-10T10:23:47.318 INFO:teuthology.orchestra.run.vm01.stdout:55834574861 2026-03-10T10:23:47.318 INFO:teuthology.orchestra.run.vm01.stdout:180388626435 2026-03-10T10:23:47.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:47 vm08 ceph-mon[50506]: pgmap v80: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 46 KiB/s, 0 objects/s recovering 2026-03-10T10:23:47.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:47 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/1791560253' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-10T10:23:47.401 INFO:tasks.cephadm.ceph_manager.ceph:need seq 55834574861 got 55834574861 for osd.1 2026-03-10T10:23:47.401 DEBUG:teuthology.parallel:result is None 2026-03-10T10:23:47.517 INFO:tasks.cephadm.ceph_manager.ceph:need seq 180388626435 got 180388626435 for osd.7 2026-03-10T10:23:47.517 INFO:teuthology.orchestra.run.vm01.stdout:115964117000 2026-03-10T10:23:47.517 DEBUG:teuthology.parallel:result is None 2026-03-10T10:23:47.587 INFO:tasks.cephadm.ceph_manager.ceph:need seq 115964117000 got 115964117000 for osd.4 2026-03-10T10:23:47.587 DEBUG:teuthology.parallel:result is None 2026-03-10T10:23:47.646 INFO:teuthology.orchestra.run.vm01.stdout:38654705679 2026-03-10T10:23:47.656 INFO:teuthology.orchestra.run.vm01.stdout:73014444043 2026-03-10T10:23:47.718 INFO:teuthology.orchestra.run.vm01.stdout:158913789957 2026-03-10T10:23:47.781 INFO:tasks.cephadm.ceph_manager.ceph:need seq 73014444043 got 73014444043 for osd.2 2026-03-10T10:23:47.781 DEBUG:teuthology.parallel:result is None 2026-03-10T10:23:47.782 INFO:tasks.cephadm.ceph_manager.ceph:need seq 38654705678 got 38654705679 for osd.0 2026-03-10T10:23:47.782 DEBUG:teuthology.parallel:result is None 2026-03-10T10:23:47.785 INFO:teuthology.orchestra.run.vm01.stdout:137438953478 2026-03-10T10:23:47.828 INFO:tasks.cephadm.ceph_manager.ceph:need seq 158913789956 got 158913789957 for osd.6 2026-03-10T10:23:47.828 DEBUG:teuthology.parallel:result is None 2026-03-10T10:23:47.866 INFO:tasks.cephadm.ceph_manager.ceph:need seq 137438953478 got 137438953478 for osd.5 2026-03-10T10:23:47.866 DEBUG:teuthology.parallel:result is None 2026-03-10T10:23:47.866 INFO:tasks.cephadm.ceph_manager.ceph:waiting for clean 2026-03-10T10:23:47.866 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph pg dump --format=json 2026-03-10T10:23:48.046 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a6b3e19a-1c6a-11f1-aaba-ff800f171c2f/mon.vm01/config 2026-03-10T10:23:48.066 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:48 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/551347168' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T10:23:48.066 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:48 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/986043752' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-10T10:23:48.066 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:48 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/2169615481' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-10T10:23:48.066 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:48 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/3083747083' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T10:23:48.066 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:48 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/574139718' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-10T10:23:48.066 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:48 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/2053079596' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-10T10:23:48.066 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:48 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/945992105' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-10T10:23:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:48 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/551347168' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T10:23:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:48 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/986043752' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-10T10:23:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:48 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/2169615481' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-10T10:23:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:48 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/3083747083' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T10:23:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:48 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/574139718' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-10T10:23:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:48 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/2053079596' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-10T10:23:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:48 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/945992105' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-10T10:23:48.374 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T10:23:48.377 INFO:teuthology.orchestra.run.vm01.stderr:dumped all 2026-03-10T10:23:48.443 INFO:teuthology.orchestra.run.vm01.stdout:{"pg_ready":true,"pg_map":{"version":81,"stamp":"2026-03-10T10:23:48.047846+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":96,"num_read_kb":82,"num_write":113,"num_write_kb":1372,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":397840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":76,"ondisk_log_size":76,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":5,"num_osds":8,"num_per_pool_osds":5,"num_per_pool_omap_osds":5,"kb":167739392,"kb_used":48692,"kb_used_data":4596,"kb_used_omap":0,"kb_used_meta":44032,"kb_avail":167690700,"statfs":{"total":171765137408,"available":171715276800,"internally_reserved":0,"allocated":4706304,"data_stored":2623931,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":45088768},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":-1,"num_objects_missing":-1,"num_objects_degraded":0,"num_objects_misplaced":-1,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":1,"num_bytes_recovered":327680,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"8.007818"},"pg_stats":[{"pgid":"1.0","version":"20'76","reported_seq":17,"reported_epoch":44,"state":"active+clean","last_fresh":"2026-03-10T10:23:41.357733+0000","last_change":"2026-03-10T10:23:41.357733+0000","last_active":"2026-03-10T10:23:41.357733+0000","last_peered":"2026-03-10T10:23:41.357733+0000","last_clean":"2026-03-10T10:23:41.357733+0000","last_became_active":"2026-03-10T10:23:41.051686+0000","last_became_peered":"2026-03-10T10:23:41.051686+0000","last_unstale":"2026-03-10T10:23:41.357733+0000","last_undegraded":"2026-03-10T10:23:41.357733+0000","last_fullsized":"2026-03-10T10:23:41.357733+0000","mapping_epoch":43,"log_start":"0'0","ondisk_log_start":"0'0","created":18,"last_epoch_clean":44,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T10:22:58.952852+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T10:22:58.952852+0000","last_clean_scrub_stamp":"2026-03-10T10:22:58.952852+0000","objects_scrubbed":0,"log_size":76,"ondisk_log_size":76,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T16:21:29.906252+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":96,"num_read_kb":82,"num_write":113,"num_write_kb":1372,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":397840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,0,6],"acting":[7,0,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":96,"num_read_kb":82,"num_write":113,"num_write_kb":1372,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":397840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1204224,"data_stored":1193520,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":76,"ondisk_log_size":76,"up":3,"acting":3,"num_store_stats":7}],"osd_stats":[{"osd":7,"up_from":42,"seq":180388626435,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6076,"kb_used_data":820,"kb_used_omap":0,"kb_used_meta":5248,"kb_avail":20961348,"statfs":{"total":21470642176,"available":21464420352,"internally_reserved":0,"allocated":839680,"data_stored":576663,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5373952},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.83299999999999996}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.85899999999999999}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.69799999999999995}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.70799999999999996}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.63800000000000001}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.84799999999999998}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":1.012}]}]},{"osd":6,"up_from":37,"seq":158913789957,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6076,"kb_used_data":820,"kb_used_omap":0,"kb_used_meta":5248,"kb_avail":20961348,"statfs":{"total":21470642176,"available":21464420352,"internally_reserved":0,"allocated":839680,"data_stored":576663,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5373952},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.34799999999999998}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.312}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.69799999999999995}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.64900000000000002}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.63200000000000001}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.61899999999999999}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.78100000000000003}]}]},{"osd":1,"up_from":13,"seq":55834574861,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6256,"kb_used_data":424,"kb_used_omap":0,"kb_used_meta":5824,"kb_avail":20961168,"statfs":{"total":21470642176,"available":21464236032,"internally_reserved":0,"allocated":434176,"data_stored":178579,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5963776},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.61499999999999999}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.65000000000000002}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.65800000000000003}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.60199999999999998}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.66800000000000004}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.63900000000000001}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.71599999999999997}]}]},{"osd":0,"up_from":9,"seq":38654705679,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6720,"kb_used_data":824,"kb_used_omap":0,"kb_used_meta":5888,"kb_avail":20960704,"statfs":{"total":21470642176,"available":21463760896,"internally_reserved":0,"allocated":843776,"data_stored":576978,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":6029312},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.33100000000000002}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.34899999999999998}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":2.1110000000000002}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.443}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.67400000000000004}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.433}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.68100000000000005}]}]},{"osd":2,"up_from":17,"seq":73014444043,"num_pgs":0,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":6260,"kb_used_data":428,"kb_used_omap":0,"kb_used_meta":5824,"kb_avail":20961164,"statfs":{"total":21470642176,"available":21464231936,"internally_reserved":0,"allocated":438272,"data_stored":178823,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5963776},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.67800000000000005}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.62}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.53700000000000003}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.64800000000000002}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.71899999999999997}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.79500000000000004}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.85199999999999998}]}]},{"osd":3,"up_from":23,"seq":98784247818,"num_pgs":0,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":5748,"kb_used_data":428,"kb_used_omap":0,"kb_used_meta":5312,"kb_avail":20961676,"statfs":{"total":21470642176,"available":21464756224,"internally_reserved":0,"allocated":438272,"data_stored":178894,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5439488},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.56100000000000005}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.371}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.52200000000000002}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.57599999999999996}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.53100000000000003}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.58499999999999996}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.68700000000000006}]}]},{"osd":4,"up_from":27,"seq":115964117000,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":5812,"kb_used_data":428,"kb_used_omap":0,"kb_used_meta":5376,"kb_avail":20961612,"statfs":{"total":21470642176,"available":21464690688,"internally_reserved":0,"allocated":438272,"data_stored":178823,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5505024},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.82999999999999996}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.69299999999999995}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.35599999999999998}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.51800000000000002}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.51100000000000001}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.499}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.73999999999999999}]}]},{"osd":5,"up_from":32,"seq":137438953479,"num_pgs":0,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":5744,"kb_used_data":424,"kb_used_omap":0,"kb_used_meta":5312,"kb_avail":20961680,"statfs":{"total":21470642176,"available":21464760320,"internally_reserved":0,"allocated":434176,"data_stored":178508,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5439488},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.874}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.71199999999999997}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.86499999999999999}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.85599999999999998}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.36899999999999999}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.81399999999999995}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.72099999999999997}]}]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-10T10:23:48.443 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph pg dump --format=json 2026-03-10T10:23:48.581 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a6b3e19a-1c6a-11f1-aaba-ff800f171c2f/mon.vm01/config 2026-03-10T10:23:48.885 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T10:23:48.886 INFO:teuthology.orchestra.run.vm01.stderr:dumped all 2026-03-10T10:23:48.946 INFO:teuthology.orchestra.run.vm01.stdout:{"pg_ready":true,"pg_map":{"version":81,"stamp":"2026-03-10T10:23:48.047846+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":96,"num_read_kb":82,"num_write":113,"num_write_kb":1372,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":397840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":76,"ondisk_log_size":76,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":5,"num_osds":8,"num_per_pool_osds":5,"num_per_pool_omap_osds":5,"kb":167739392,"kb_used":48692,"kb_used_data":4596,"kb_used_omap":0,"kb_used_meta":44032,"kb_avail":167690700,"statfs":{"total":171765137408,"available":171715276800,"internally_reserved":0,"allocated":4706304,"data_stored":2623931,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":45088768},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":-1,"num_objects_missing":-1,"num_objects_degraded":0,"num_objects_misplaced":-1,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":1,"num_bytes_recovered":327680,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"8.007818"},"pg_stats":[{"pgid":"1.0","version":"20'76","reported_seq":17,"reported_epoch":44,"state":"active+clean","last_fresh":"2026-03-10T10:23:41.357733+0000","last_change":"2026-03-10T10:23:41.357733+0000","last_active":"2026-03-10T10:23:41.357733+0000","last_peered":"2026-03-10T10:23:41.357733+0000","last_clean":"2026-03-10T10:23:41.357733+0000","last_became_active":"2026-03-10T10:23:41.051686+0000","last_became_peered":"2026-03-10T10:23:41.051686+0000","last_unstale":"2026-03-10T10:23:41.357733+0000","last_undegraded":"2026-03-10T10:23:41.357733+0000","last_fullsized":"2026-03-10T10:23:41.357733+0000","mapping_epoch":43,"log_start":"0'0","ondisk_log_start":"0'0","created":18,"last_epoch_clean":44,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T10:22:58.952852+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T10:22:58.952852+0000","last_clean_scrub_stamp":"2026-03-10T10:22:58.952852+0000","objects_scrubbed":0,"log_size":76,"ondisk_log_size":76,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T16:21:29.906252+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":96,"num_read_kb":82,"num_write":113,"num_write_kb":1372,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":397840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,0,6],"acting":[7,0,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":96,"num_read_kb":82,"num_write":113,"num_write_kb":1372,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":397840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1204224,"data_stored":1193520,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":76,"ondisk_log_size":76,"up":3,"acting":3,"num_store_stats":7}],"osd_stats":[{"osd":7,"up_from":42,"seq":180388626435,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6076,"kb_used_data":820,"kb_used_omap":0,"kb_used_meta":5248,"kb_avail":20961348,"statfs":{"total":21470642176,"available":21464420352,"internally_reserved":0,"allocated":839680,"data_stored":576663,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5373952},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.83299999999999996}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.85899999999999999}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.69799999999999995}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.70799999999999996}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.63800000000000001}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.84799999999999998}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":1.012}]}]},{"osd":6,"up_from":37,"seq":158913789957,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6076,"kb_used_data":820,"kb_used_omap":0,"kb_used_meta":5248,"kb_avail":20961348,"statfs":{"total":21470642176,"available":21464420352,"internally_reserved":0,"allocated":839680,"data_stored":576663,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5373952},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.34799999999999998}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.312}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.69799999999999995}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.64900000000000002}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.63200000000000001}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.61899999999999999}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.78100000000000003}]}]},{"osd":1,"up_from":13,"seq":55834574861,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6256,"kb_used_data":424,"kb_used_omap":0,"kb_used_meta":5824,"kb_avail":20961168,"statfs":{"total":21470642176,"available":21464236032,"internally_reserved":0,"allocated":434176,"data_stored":178579,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5963776},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.61499999999999999}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.65000000000000002}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.65800000000000003}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.60199999999999998}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.66800000000000004}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.63900000000000001}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.71599999999999997}]}]},{"osd":0,"up_from":9,"seq":38654705679,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6720,"kb_used_data":824,"kb_used_omap":0,"kb_used_meta":5888,"kb_avail":20960704,"statfs":{"total":21470642176,"available":21463760896,"internally_reserved":0,"allocated":843776,"data_stored":576978,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":6029312},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.33100000000000002}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.34899999999999998}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":2.1110000000000002}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.443}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.67400000000000004}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.433}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.68100000000000005}]}]},{"osd":2,"up_from":17,"seq":73014444043,"num_pgs":0,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":6260,"kb_used_data":428,"kb_used_omap":0,"kb_used_meta":5824,"kb_avail":20961164,"statfs":{"total":21470642176,"available":21464231936,"internally_reserved":0,"allocated":438272,"data_stored":178823,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5963776},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.67800000000000005}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.62}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.53700000000000003}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.64800000000000002}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.71899999999999997}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.79500000000000004}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.85199999999999998}]}]},{"osd":3,"up_from":23,"seq":98784247818,"num_pgs":0,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":5748,"kb_used_data":428,"kb_used_omap":0,"kb_used_meta":5312,"kb_avail":20961676,"statfs":{"total":21470642176,"available":21464756224,"internally_reserved":0,"allocated":438272,"data_stored":178894,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5439488},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.56100000000000005}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.371}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.52200000000000002}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.57599999999999996}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.53100000000000003}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.58499999999999996}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.68700000000000006}]}]},{"osd":4,"up_from":27,"seq":115964117000,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":5812,"kb_used_data":428,"kb_used_omap":0,"kb_used_meta":5376,"kb_avail":20961612,"statfs":{"total":21470642176,"available":21464690688,"internally_reserved":0,"allocated":438272,"data_stored":178823,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5505024},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.82999999999999996}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.69299999999999995}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.35599999999999998}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.51800000000000002}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.51100000000000001}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.499}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.73999999999999999}]}]},{"osd":5,"up_from":32,"seq":137438953479,"num_pgs":0,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":5744,"kb_used_data":424,"kb_used_omap":0,"kb_used_meta":5312,"kb_avail":20961680,"statfs":{"total":21470642176,"available":21464760320,"internally_reserved":0,"allocated":434176,"data_stored":178508,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5439488},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.874}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.71199999999999997}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.86499999999999999}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.85599999999999998}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.36899999999999999}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.81399999999999995}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.72099999999999997}]}]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-10T10:23:48.947 INFO:tasks.cephadm.ceph_manager.ceph:clean! 2026-03-10T10:23:48.947 INFO:tasks.ceph:Waiting until ceph cluster ceph is healthy... 2026-03-10T10:23:48.947 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy 2026-03-10T10:23:48.947 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph health --format=json 2026-03-10T10:23:49.087 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a6b3e19a-1c6a-11f1-aaba-ff800f171c2f/mon.vm01/config 2026-03-10T10:23:49.135 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:49 vm01 ceph-mon[48597]: pgmap v81: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 40 KiB/s, 0 objects/s recovering 2026-03-10T10:23:49.135 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:49 vm01 ceph-mon[48597]: from='client.14462 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:23:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:49 vm08 ceph-mon[50506]: pgmap v81: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 40 KiB/s, 0 objects/s recovering 2026-03-10T10:23:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:49 vm08 ceph-mon[50506]: from='client.14462 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:23:49.415 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T10:23:49.415 INFO:teuthology.orchestra.run.vm01.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-03-10T10:23:49.457 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy done 2026-03-10T10:23:49.457 INFO:tasks.cephadm:Setup complete, yielding 2026-03-10T10:23:49.457 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-10T10:23:49.459 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm01.local 2026-03-10T10:23:49.459 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- bash -c 'ceph orch status' 2026-03-10T10:23:49.895 INFO:teuthology.orchestra.run.vm01.stdout:Backend: cephadm 2026-03-10T10:23:49.895 INFO:teuthology.orchestra.run.vm01.stdout:Available: Yes 2026-03-10T10:23:49.895 INFO:teuthology.orchestra.run.vm01.stdout:Paused: No 2026-03-10T10:23:49.957 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- bash -c 'ceph orch ps' 2026-03-10T10:23:50.200 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:50 vm01 ceph-mon[48597]: from='client.14466 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:23:50.200 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:50 vm01 ceph-mon[48597]: from='client.? 192.168.123.101:0/3774416212' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-10T10:23:50.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:50 vm08 ceph-mon[50506]: from='client.14466 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:23:50.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:50 vm08 ceph-mon[50506]: from='client.? 192.168.123.101:0/3774416212' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-10T10:23:50.418 INFO:teuthology.orchestra.run.vm01.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-10T10:23:50.418 INFO:teuthology.orchestra.run.vm01.stdout:alertmanager.vm01 vm01 *:9093,9094 running (76s) 46s ago 2m 16.0M - ba2b418f427c 7ff797f1c9c6 2026-03-10T10:23:50.418 INFO:teuthology.orchestra.run.vm01.stdout:crash.vm01 vm01 running (2m) 46s ago 2m 6945k - 17.2.0 e1d6a67b021e 6afeb9006343 2026-03-10T10:23:50.418 INFO:teuthology.orchestra.run.vm01.stdout:crash.vm08 vm08 running (96s) 13s ago 96s 7101k - 17.2.0 e1d6a67b021e 7cda24507e64 2026-03-10T10:23:50.418 INFO:teuthology.orchestra.run.vm01.stdout:grafana.vm01 vm01 *:3000 running (73s) 46s ago 108s 42.5M - 8.3.5 dad864ee21e9 34d30cb61a3d 2026-03-10T10:23:50.418 INFO:teuthology.orchestra.run.vm01.stdout:mgr.vm01.nkkzmc vm01 *:9283 running (2m) 46s ago 2m 457M - 17.2.0 e1d6a67b021e 10df7af51a16 2026-03-10T10:23:50.418 INFO:teuthology.orchestra.run.vm01.stdout:mgr.vm08.rwpjaf vm08 *:8443,9283 running (95s) 13s ago 95s 419M - 17.2.0 e1d6a67b021e 879cc98066f9 2026-03-10T10:23:50.418 INFO:teuthology.orchestra.run.vm01.stdout:mon.vm01 vm01 running (2m) 46s ago 2m 46.0M 2048M 17.2.0 e1d6a67b021e ab7750cedb59 2026-03-10T10:23:50.418 INFO:teuthology.orchestra.run.vm01.stdout:mon.vm08 vm08 running (94s) 13s ago 94s 40.9M 2048M 17.2.0 e1d6a67b021e 885cea287d64 2026-03-10T10:23:50.418 INFO:teuthology.orchestra.run.vm01.stdout:node-exporter.vm01 vm01 *:9100 running (106s) 46s ago 106s 18.4M - 1dbe0e931976 6c2f4d6573ea 2026-03-10T10:23:50.418 INFO:teuthology.orchestra.run.vm01.stdout:node-exporter.vm08 vm08 *:9100 running (90s) 13s ago 90s 15.9M - 1dbe0e931976 b57df444df6c 2026-03-10T10:23:50.418 INFO:teuthology.orchestra.run.vm01.stdout:osd.0 vm01 running (74s) 46s ago 74s 32.1M 4096M 17.2.0 e1d6a67b021e 85d0fda3e6e6 2026-03-10T10:23:50.418 INFO:teuthology.orchestra.run.vm01.stdout:osd.1 vm01 running (65s) 46s ago 65s 36.9M 4096M 17.2.0 e1d6a67b021e ba9f1f6a4878 2026-03-10T10:23:50.418 INFO:teuthology.orchestra.run.vm01.stdout:osd.2 vm01 running (56s) 46s ago 56s 34.7M 4096M 17.2.0 e1d6a67b021e c28c1b8350bb 2026-03-10T10:23:50.418 INFO:teuthology.orchestra.run.vm01.stdout:osd.3 vm01 running (47s) 46s ago 47s 11.8M 4096M 17.2.0 e1d6a67b021e ad489b45223c 2026-03-10T10:23:50.418 INFO:teuthology.orchestra.run.vm01.stdout:osd.4 vm08 running (39s) 13s ago 39s 35.4M 4096M 17.2.0 e1d6a67b021e e9bde1672bf0 2026-03-10T10:23:50.418 INFO:teuthology.orchestra.run.vm01.stdout:osd.5 vm08 running (32s) 13s ago 32s 31.4M 4096M 17.2.0 e1d6a67b021e b862891b87e0 2026-03-10T10:23:50.418 INFO:teuthology.orchestra.run.vm01.stdout:osd.6 vm08 running (24s) 13s ago 24s 32.0M 4096M 17.2.0 e1d6a67b021e cbab1f52a0c0 2026-03-10T10:23:50.418 INFO:teuthology.orchestra.run.vm01.stdout:osd.7 vm08 running (15s) 13s ago 15s 15.4M 4096M 17.2.0 e1d6a67b021e cb553cf993cf 2026-03-10T10:23:50.418 INFO:teuthology.orchestra.run.vm01.stdout:prometheus.vm01 vm01 *:9095 running (83s) 46s ago 83s 34.5M - 514e6a882f6e f370820f31be 2026-03-10T10:23:50.482 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- bash -c 'ceph orch ls' 2026-03-10T10:23:50.918 INFO:teuthology.orchestra.run.vm01.stdout:NAME PORTS RUNNING REFRESHED AGE PLACEMENT 2026-03-10T10:23:50.918 INFO:teuthology.orchestra.run.vm01.stdout:alertmanager ?:9093,9094 1/1 46s ago 2m count:1 2026-03-10T10:23:50.918 INFO:teuthology.orchestra.run.vm01.stdout:crash 2/2 46s ago 2m * 2026-03-10T10:23:50.918 INFO:teuthology.orchestra.run.vm01.stdout:grafana ?:3000 1/1 46s ago 2m count:1 2026-03-10T10:23:50.918 INFO:teuthology.orchestra.run.vm01.stdout:mgr 2/2 46s ago 2m count:2 2026-03-10T10:23:50.918 INFO:teuthology.orchestra.run.vm01.stdout:mon 2/2 46s ago 2m vm01:192.168.123.101=vm01;vm08:192.168.123.108=vm08;count:2 2026-03-10T10:23:50.918 INFO:teuthology.orchestra.run.vm01.stdout:node-exporter ?:9100 2/2 46s ago 2m * 2026-03-10T10:23:50.918 INFO:teuthology.orchestra.run.vm01.stdout:osd 8 46s ago - 2026-03-10T10:23:50.918 INFO:teuthology.orchestra.run.vm01.stdout:prometheus ?:9095 1/1 46s ago 2m count:1 2026-03-10T10:23:50.983 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- bash -c 'ceph orch host ls' 2026-03-10T10:23:51.226 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:51 vm01 ceph-mon[48597]: from='client.14474 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:23:51.226 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:51 vm01 ceph-mon[48597]: pgmap v82: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 32 KiB/s, 0 objects/s recovering 2026-03-10T10:23:51.226 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:51 vm01 ceph-mon[48597]: from='client.14478 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:23:51.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:51 vm08 ceph-mon[50506]: from='client.14474 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:23:51.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:51 vm08 ceph-mon[50506]: pgmap v82: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 32 KiB/s, 0 objects/s recovering 2026-03-10T10:23:51.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:51 vm08 ceph-mon[50506]: from='client.14478 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:23:51.428 INFO:teuthology.orchestra.run.vm01.stdout:HOST ADDR LABELS STATUS 2026-03-10T10:23:51.428 INFO:teuthology.orchestra.run.vm01.stdout:vm01 192.168.123.101 2026-03-10T10:23:51.428 INFO:teuthology.orchestra.run.vm01.stdout:vm08 192.168.123.108 2026-03-10T10:23:51.428 INFO:teuthology.orchestra.run.vm01.stdout:2 hosts in cluster 2026-03-10T10:23:51.472 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- bash -c 'ceph orch device ls' 2026-03-10T10:23:51.900 INFO:teuthology.orchestra.run.vm01.stdout:HOST PATH TYPE DEVICE ID SIZE AVAILABLE REJECT REASONS 2026-03-10T10:23:51.901 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/vdb hdd DWNBRSTVMM01001 21.4G Insufficient space (<10 extents) on vgs, LVM detected, locked 2026-03-10T10:23:51.901 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/vdc hdd DWNBRSTVMM01002 21.4G Insufficient space (<10 extents) on vgs, LVM detected, locked 2026-03-10T10:23:51.901 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/vdd hdd DWNBRSTVMM01003 21.4G Insufficient space (<10 extents) on vgs, LVM detected, locked 2026-03-10T10:23:51.901 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/vde hdd DWNBRSTVMM01004 21.4G Insufficient space (<10 extents) on vgs, LVM detected, locked 2026-03-10T10:23:51.901 INFO:teuthology.orchestra.run.vm01.stdout:vm08 /dev/vdb hdd DWNBRSTVMM08001 21.4G Insufficient space (<10 extents) on vgs, LVM detected, locked 2026-03-10T10:23:51.901 INFO:teuthology.orchestra.run.vm01.stdout:vm08 /dev/vdc hdd DWNBRSTVMM08002 21.4G Insufficient space (<10 extents) on vgs, LVM detected, locked 2026-03-10T10:23:51.901 INFO:teuthology.orchestra.run.vm01.stdout:vm08 /dev/vdd hdd DWNBRSTVMM08003 21.4G Insufficient space (<10 extents) on vgs, LVM detected, locked 2026-03-10T10:23:51.901 INFO:teuthology.orchestra.run.vm01.stdout:vm08 /dev/vde hdd DWNBRSTVMM08004 21.4G Insufficient space (<10 extents) on vgs, LVM detected, locked 2026-03-10T10:23:51.966 INFO:teuthology.run_tasks:Running task vip.exec... 2026-03-10T10:23:51.969 INFO:tasks.vip:Running commands on role host.a host ubuntu@vm01.local 2026-03-10T10:23:51.969 DEBUG:teuthology.orchestra.run.vm01:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c 'systemctl stop nfs-server' 2026-03-10T10:23:51.992 INFO:teuthology.orchestra.run.vm01.stderr:+ systemctl stop nfs-server 2026-03-10T10:23:51.999 INFO:tasks.vip:Running commands on role host.b host ubuntu@vm08.local 2026-03-10T10:23:51.999 DEBUG:teuthology.orchestra.run.vm08:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c 'systemctl stop nfs-server' 2026-03-10T10:23:52.026 INFO:teuthology.orchestra.run.vm08.stderr:+ systemctl stop nfs-server 2026-03-10T10:23:52.032 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-10T10:23:52.035 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm01.local 2026-03-10T10:23:52.035 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- bash -c 'ceph fs volume create foofs' 2026-03-10T10:23:52.313 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:52 vm01 ceph-mon[48597]: from='client.14482 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:23:52.313 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:52 vm01 ceph-mon[48597]: from='client.14486 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:23:52.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:52 vm08 ceph-mon[50506]: from='client.14482 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:23:52.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:52 vm08 ceph-mon[50506]: from='client.14486 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:23:53.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:53 vm08 ceph-mon[50506]: from='client.14490 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:23:53.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:53 vm08 ceph-mon[50506]: pgmap v83: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 29 KiB/s, 0 objects/s recovering 2026-03-10T10:23:53.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:53 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd pool create", "pool": "cephfs.foofs.meta"}]: dispatch 2026-03-10T10:23:53.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:53 vm01 ceph-mon[48597]: from='client.14490 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:23:53.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:53 vm01 ceph-mon[48597]: pgmap v83: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 29 KiB/s, 0 objects/s recovering 2026-03-10T10:23:53.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:53 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd pool create", "pool": "cephfs.foofs.meta"}]: dispatch 2026-03-10T10:23:54.195 INFO:teuthology.run_tasks:Running task cephadm.wait_for_service... 2026-03-10T10:23:54.198 INFO:tasks.cephadm:Waiting for ceph service mds.foofs to start (timeout 300)... 2026-03-10T10:23:54.198 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph orch ls -f json 2026-03-10T10:23:54.375 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:54 vm01 ceph-a6b3e19a-1c6a-11f1-aaba-ff800f171c2f-mon-vm01[48593]: 2026-03-10T10:23:54.117+0000 7fb0c5228700 -1 log_channel(cluster) log [ERR] : Health check failed: 1 filesystem is offline (MDS_ALL_DOWN) 2026-03-10T10:23:54.375 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:54 vm01 ceph-mon[48597]: from='client.24309 -' entity='client.admin' cmd=[{"prefix": "fs volume create", "name": "foofs", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:23:54.375 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:54 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "osd pool create", "pool": "cephfs.foofs.meta"}]': finished 2026-03-10T10:23:54.375 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:54 vm01 ceph-mon[48597]: osdmap e45: 8 total, 8 up, 8 in 2026-03-10T10:23:54.375 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:54 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd pool create", "pool": "cephfs.foofs.data"}]: dispatch 2026-03-10T10:23:54.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:54 vm08 ceph-mon[50506]: from='client.24309 -' entity='client.admin' cmd=[{"prefix": "fs volume create", "name": "foofs", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:23:54.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:54 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "osd pool create", "pool": "cephfs.foofs.meta"}]': finished 2026-03-10T10:23:54.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:54 vm08 ceph-mon[50506]: osdmap e45: 8 total, 8 up, 8 in 2026-03-10T10:23:54.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:54 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd pool create", "pool": "cephfs.foofs.data"}]: dispatch 2026-03-10T10:23:54.749 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T10:23:54.749 INFO:teuthology.orchestra.run.vm01.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-10T10:21:32.962198Z", "last_refresh": "2026-03-10T10:23:03.968119Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-10T10:21:31.589340Z", "last_refresh": "2026-03-10T10:23:03.968159Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "status": {"created": "2026-03-10T10:21:32.232037Z", "last_refresh": "2026-03-10T10:23:03.968188Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:23:54.145652Z service:mds.foofs [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "foofs", "service_name": "mds.foofs", "service_type": "mds", "status": {"created": "2026-03-10T10:23:54.135972Z", "running": 0, "size": 2}}, {"placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-10T10:21:31.311502Z", "last_refresh": "2026-03-10T10:23:03.968071Z", "running": 2, "size": 2}}, {"placement": {"count": 2, "hosts": ["vm01:192.168.123.101=vm01", "vm08:192.168.123.108=vm08"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-10T10:21:48.995982Z", "last_refresh": "2026-03-10T10:23:03.967954Z", "running": 2, "size": 2}}, {"placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-10T10:21:32.604681Z", "last_refresh": "2026-03-10T10:23:03.968215Z", "ports": [9100], "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "e1d6a67b021eb077ee22bf650f1a9fb1980a2cf5c36bdb9cba9eac6de8f702d9", "container_image_name": "quay.io/ceph/ceph@sha256:12a0a4f43413fd97a14a3d47a3451b2d2df50020835bb93db666209f3f77617a", "last_refresh": "2026-03-10T10:23:03.968270Z", "running": 8, "size": 8}, "unmanaged": true}, {"placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-10T10:21:31.900421Z", "last_refresh": "2026-03-10T10:23:03.968243Z", "ports": [9095], "running": 1, "size": 1}}] 2026-03-10T10:23:54.842 INFO:tasks.cephadm:mds.foofs has 0/2 2026-03-10T10:23:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:55 vm01 ceph-mon[48597]: pgmap v85: 33 pgs: 10 creating+peering, 1 active+clean, 22 unknown; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 32 KiB/s, 0 objects/s recovering 2026-03-10T10:23:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:55 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "osd pool create", "pool": "cephfs.foofs.data"}]': finished 2026-03-10T10:23:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:55 vm01 ceph-mon[48597]: osdmap e46: 8 total, 8 up, 8 in 2026-03-10T10:23:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:55 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "fs new", "fs_name": "foofs", "metadata": "cephfs.foofs.meta", "data": "cephfs.foofs.data"}]: dispatch 2026-03-10T10:23:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:55 vm01 ceph-mon[48597]: Health check failed: 1 filesystem is offline (MDS_ALL_DOWN) 2026-03-10T10:23:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:55 vm01 ceph-mon[48597]: Health check failed: 1 filesystem is online with fewer MDS than max_mds (MDS_UP_LESS_THAN_MAX) 2026-03-10T10:23:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:55 vm01 ceph-mon[48597]: osdmap e47: 8 total, 8 up, 8 in 2026-03-10T10:23:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:55 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "fs new", "fs_name": "foofs", "metadata": "cephfs.foofs.meta", "data": "cephfs.foofs.data"}]': finished 2026-03-10T10:23:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:55 vm01 ceph-mon[48597]: fsmap foofs:0 2026-03-10T10:23:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:55 vm01 ceph-mon[48597]: Saving service mds.foofs spec with placement count:2 2026-03-10T10:23:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:55 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:55 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:23:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:55 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:55 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:23:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:55 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:55 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:55 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "mds.foofs.vm01.lfesqj", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]: dispatch 2026-03-10T10:23:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:55 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "auth get-or-create", "entity": "mds.foofs.vm01.lfesqj", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]': finished 2026-03-10T10:23:55.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:55 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:55.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:55 vm01 ceph-mon[48597]: Deploying daemon mds.foofs.vm01.lfesqj on vm01 2026-03-10T10:23:55.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:55 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:55.422 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:55 vm08 ceph-mon[50506]: pgmap v85: 33 pgs: 10 creating+peering, 1 active+clean, 22 unknown; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 32 KiB/s, 0 objects/s recovering 2026-03-10T10:23:55.422 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:55 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "osd pool create", "pool": "cephfs.foofs.data"}]': finished 2026-03-10T10:23:55.422 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:55 vm08 ceph-mon[50506]: osdmap e46: 8 total, 8 up, 8 in 2026-03-10T10:23:55.422 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:55 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "fs new", "fs_name": "foofs", "metadata": "cephfs.foofs.meta", "data": "cephfs.foofs.data"}]: dispatch 2026-03-10T10:23:55.422 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:55 vm08 ceph-mon[50506]: Health check failed: 1 filesystem is offline (MDS_ALL_DOWN) 2026-03-10T10:23:55.422 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:55 vm08 ceph-mon[50506]: Health check failed: 1 filesystem is online with fewer MDS than max_mds (MDS_UP_LESS_THAN_MAX) 2026-03-10T10:23:55.422 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:55 vm08 ceph-mon[50506]: osdmap e47: 8 total, 8 up, 8 in 2026-03-10T10:23:55.422 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:55 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "fs new", "fs_name": "foofs", "metadata": "cephfs.foofs.meta", "data": "cephfs.foofs.data"}]': finished 2026-03-10T10:23:55.422 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:55 vm08 ceph-mon[50506]: fsmap foofs:0 2026-03-10T10:23:55.422 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:55 vm08 ceph-mon[50506]: Saving service mds.foofs spec with placement count:2 2026-03-10T10:23:55.422 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:55 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:55.422 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:55 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:23:55.422 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:55 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:55.422 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:55 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:23:55.422 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:55 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:55.422 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:55 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:55.422 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:55 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "mds.foofs.vm01.lfesqj", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]: dispatch 2026-03-10T10:23:55.422 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:55 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "auth get-or-create", "entity": "mds.foofs.vm01.lfesqj", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]': finished 2026-03-10T10:23:55.422 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:55 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:55.422 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:55 vm08 ceph-mon[50506]: Deploying daemon mds.foofs.vm01.lfesqj on vm01 2026-03-10T10:23:55.422 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:55 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:55.843 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph orch ls -f json 2026-03-10T10:23:56.475 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:56 vm01 ceph-mon[48597]: from='client.14498 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:23:56.475 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:56 vm01 ceph-mon[48597]: osdmap e48: 8 total, 8 up, 8 in 2026-03-10T10:23:56.475 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:56 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:56.475 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:56 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "mds.foofs.vm08.xjgpjf", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]: dispatch 2026-03-10T10:23:56.475 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:56 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "auth get-or-create", "entity": "mds.foofs.vm08.xjgpjf", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]': finished 2026-03-10T10:23:56.475 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:56 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:56.475 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:56 vm01 ceph-mon[48597]: Deploying daemon mds.foofs.vm08.xjgpjf on vm08 2026-03-10T10:23:56.475 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:56 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:56.475 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:56 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:23:56.475 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:56 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:56.475 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:56 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:23:56.475 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T10:23:56.475 INFO:teuthology.orchestra.run.vm01.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-10T10:21:32.962198Z", "last_refresh": "2026-03-10T10:23:03.968119Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-10T10:21:31.589340Z", "last_refresh": "2026-03-10T10:23:03.968159Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "status": {"created": "2026-03-10T10:21:32.232037Z", "last_refresh": "2026-03-10T10:23:03.968188Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:23:54.145652Z service:mds.foofs [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "foofs", "service_name": "mds.foofs", "service_type": "mds", "status": {"created": "2026-03-10T10:23:54.135972Z", "running": 0, "size": 2}}, {"placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-10T10:21:31.311502Z", "last_refresh": "2026-03-10T10:23:03.968071Z", "running": 2, "size": 2}}, {"placement": {"count": 2, "hosts": ["vm01:192.168.123.101=vm01", "vm08:192.168.123.108=vm08"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-10T10:21:48.995982Z", "last_refresh": "2026-03-10T10:23:03.967954Z", "running": 2, "size": 2}}, {"placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-10T10:21:32.604681Z", "last_refresh": "2026-03-10T10:23:03.968215Z", "ports": [9100], "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "e1d6a67b021eb077ee22bf650f1a9fb1980a2cf5c36bdb9cba9eac6de8f702d9", "container_image_name": "quay.io/ceph/ceph@sha256:12a0a4f43413fd97a14a3d47a3451b2d2df50020835bb93db666209f3f77617a", "last_refresh": "2026-03-10T10:23:03.968270Z", "running": 8, "size": 8}, "unmanaged": true}, {"placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-10T10:21:31.900421Z", "last_refresh": "2026-03-10T10:23:03.968243Z", "ports": [9095], "running": 1, "size": 1}}] 2026-03-10T10:23:56.480 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:56 vm08 ceph-mon[50506]: from='client.14498 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:23:56.480 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:56 vm08 ceph-mon[50506]: osdmap e48: 8 total, 8 up, 8 in 2026-03-10T10:23:56.480 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:56 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:56.480 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:56 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "mds.foofs.vm08.xjgpjf", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]: dispatch 2026-03-10T10:23:56.480 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:56 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "auth get-or-create", "entity": "mds.foofs.vm08.xjgpjf", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]': finished 2026-03-10T10:23:56.480 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:56 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:56.480 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:56 vm08 ceph-mon[50506]: Deploying daemon mds.foofs.vm08.xjgpjf on vm08 2026-03-10T10:23:56.480 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:56 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:56.480 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:56 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:23:56.480 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:56 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:56.480 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:56 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:23:56.535 INFO:tasks.cephadm:mds.foofs has 0/2 2026-03-10T10:23:57.300 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:57 vm01 ceph-mon[48597]: pgmap v89: 65 pgs: 33 active+clean, 10 creating+peering, 22 unknown; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:23:57.300 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:57 vm01 ceph-mon[48597]: mds.? [v2:192.168.123.101:6834/921717464,v1:192.168.123.101:6835/921717464] up:boot 2026-03-10T10:23:57.300 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:57 vm01 ceph-mon[48597]: mds.? [v2:192.168.123.108:6832/639137048,v1:192.168.123.108:6833/639137048] up:boot 2026-03-10T10:23:57.300 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:57 vm01 ceph-mon[48597]: daemon mds.foofs.vm01.lfesqj assigned to filesystem foofs as rank 0 (now has 1 ranks) 2026-03-10T10:23:57.300 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:57 vm01 ceph-mon[48597]: Health check cleared: MDS_ALL_DOWN (was: 1 filesystem is offline) 2026-03-10T10:23:57.300 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:57 vm01 ceph-mon[48597]: Health check cleared: MDS_UP_LESS_THAN_MAX (was: 1 filesystem is online with fewer MDS than max_mds) 2026-03-10T10:23:57.300 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:57 vm01 ceph-mon[48597]: Cluster is now healthy 2026-03-10T10:23:57.300 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:57 vm01 ceph-mon[48597]: fsmap foofs:0 2 up:standby 2026-03-10T10:23:57.300 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:57 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mds metadata", "who": "foofs.vm01.lfesqj"}]: dispatch 2026-03-10T10:23:57.300 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:57 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mds metadata", "who": "foofs.vm08.xjgpjf"}]: dispatch 2026-03-10T10:23:57.300 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:57 vm01 ceph-mon[48597]: fsmap foofs:1 {0=foofs.vm01.lfesqj=up:creating} 1 up:standby 2026-03-10T10:23:57.300 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:57 vm01 ceph-mon[48597]: osdmap e49: 8 total, 8 up, 8 in 2026-03-10T10:23:57.300 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:57 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mds metadata", "who": "foofs.vm01.lfesqj"}]: dispatch 2026-03-10T10:23:57.300 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:57 vm01 ceph-mon[48597]: daemon mds.foofs.vm01.lfesqj is now active in filesystem foofs as rank 0 2026-03-10T10:23:57.300 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:57 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:57.300 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:57 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:57.536 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- ceph orch ls -f json 2026-03-10T10:23:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:57 vm08 ceph-mon[50506]: pgmap v89: 65 pgs: 33 active+clean, 10 creating+peering, 22 unknown; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:23:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:57 vm08 ceph-mon[50506]: mds.? [v2:192.168.123.101:6834/921717464,v1:192.168.123.101:6835/921717464] up:boot 2026-03-10T10:23:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:57 vm08 ceph-mon[50506]: mds.? [v2:192.168.123.108:6832/639137048,v1:192.168.123.108:6833/639137048] up:boot 2026-03-10T10:23:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:57 vm08 ceph-mon[50506]: daemon mds.foofs.vm01.lfesqj assigned to filesystem foofs as rank 0 (now has 1 ranks) 2026-03-10T10:23:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:57 vm08 ceph-mon[50506]: Health check cleared: MDS_ALL_DOWN (was: 1 filesystem is offline) 2026-03-10T10:23:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:57 vm08 ceph-mon[50506]: Health check cleared: MDS_UP_LESS_THAN_MAX (was: 1 filesystem is online with fewer MDS than max_mds) 2026-03-10T10:23:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:57 vm08 ceph-mon[50506]: Cluster is now healthy 2026-03-10T10:23:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:57 vm08 ceph-mon[50506]: fsmap foofs:0 2 up:standby 2026-03-10T10:23:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:57 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mds metadata", "who": "foofs.vm01.lfesqj"}]: dispatch 2026-03-10T10:23:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:57 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mds metadata", "who": "foofs.vm08.xjgpjf"}]: dispatch 2026-03-10T10:23:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:57 vm08 ceph-mon[50506]: fsmap foofs:1 {0=foofs.vm01.lfesqj=up:creating} 1 up:standby 2026-03-10T10:23:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:57 vm08 ceph-mon[50506]: osdmap e49: 8 total, 8 up, 8 in 2026-03-10T10:23:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:57 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "mds metadata", "who": "foofs.vm01.lfesqj"}]: dispatch 2026-03-10T10:23:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:57 vm08 ceph-mon[50506]: daemon mds.foofs.vm01.lfesqj is now active in filesystem foofs as rank 0 2026-03-10T10:23:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:57 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:57 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:57.975 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T10:23:57.975 INFO:teuthology.orchestra.run.vm01.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-10T10:21:32.962198Z", "last_refresh": "2026-03-10T10:23:57.303140Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-10T10:21:31.589340Z", "last_refresh": "2026-03-10T10:23:56.758052Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "status": {"created": "2026-03-10T10:21:32.232037Z", "last_refresh": "2026-03-10T10:23:57.303201Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:23:54.145652Z service:mds.foofs [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "foofs", "service_name": "mds.foofs", "service_type": "mds", "status": {"created": "2026-03-10T10:23:54.135972Z", "last_refresh": "2026-03-10T10:23:56.758356Z", "running": 2, "size": 2}}, {"placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-10T10:21:31.311502Z", "last_refresh": "2026-03-10T10:23:56.758173Z", "running": 2, "size": 2}}, {"placement": {"count": 2, "hosts": ["vm01:192.168.123.101=vm01", "vm08:192.168.123.108=vm08"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-10T10:21:48.995982Z", "last_refresh": "2026-03-10T10:23:56.758205Z", "running": 2, "size": 2}}, {"placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-10T10:21:32.604681Z", "last_refresh": "2026-03-10T10:23:56.758232Z", "ports": [9100], "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "e1d6a67b021eb077ee22bf650f1a9fb1980a2cf5c36bdb9cba9eac6de8f702d9", "container_image_name": "quay.io/ceph/ceph@sha256:12a0a4f43413fd97a14a3d47a3451b2d2df50020835bb93db666209f3f77617a", "last_refresh": "2026-03-10T10:23:56.758258Z", "running": 8, "size": 8}, "unmanaged": true}, {"placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-10T10:21:31.900421Z", "last_refresh": "2026-03-10T10:23:57.303252Z", "ports": [9095], "running": 1, "size": 1}}] 2026-03-10T10:23:58.022 INFO:tasks.cephadm:mds.foofs has 2/2 2026-03-10T10:23:58.022 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-10T10:23:58.024 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm01.local 2026-03-10T10:23:58.024 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- bash -c 'ceph nfs cluster create foo --placement=2 || ceph nfs cluster create cephfs foo --placement=2' 2026-03-10T10:23:58.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:58 vm08 ceph-mon[50506]: from='client.14508 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:23:58.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:58 vm08 ceph-mon[50506]: mds.? [v2:192.168.123.101:6834/921717464,v1:192.168.123.101:6835/921717464] up:active 2026-03-10T10:23:58.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:58 vm08 ceph-mon[50506]: fsmap foofs:1 {0=foofs.vm01.lfesqj=up:active} 1 up:standby 2026-03-10T10:23:58.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:58 vm08 ceph-mon[50506]: fsmap foofs:1 {0=foofs.vm01.lfesqj=up:active} 1 up:standby 2026-03-10T10:23:58.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:58 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:58.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:58 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:58.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:58 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:23:58.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:58 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:58.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:58 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:23:58.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:58 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:58.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:58 vm01 ceph-mon[48597]: from='client.14508 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:23:58.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:58 vm01 ceph-mon[48597]: mds.? [v2:192.168.123.101:6834/921717464,v1:192.168.123.101:6835/921717464] up:active 2026-03-10T10:23:58.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:58 vm01 ceph-mon[48597]: fsmap foofs:1 {0=foofs.vm01.lfesqj=up:active} 1 up:standby 2026-03-10T10:23:58.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:58 vm01 ceph-mon[48597]: fsmap foofs:1 {0=foofs.vm01.lfesqj=up:active} 1 up:standby 2026-03-10T10:23:58.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:58 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:58.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:58 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:58.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:58 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:23:58.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:58 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:23:58.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:58 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:23:58.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:58 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:23:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:59 vm08 ceph-mon[50506]: from='client.14512 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:23:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:59 vm08 ceph-mon[50506]: pgmap v91: 65 pgs: 52 active+clean, 10 creating+peering, 3 unknown; 449 KiB data, 49 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:23:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:59 vm08 ceph-mon[50506]: from='client.24321 -' entity='client.admin' cmd=[{"prefix": "nfs cluster create", "cluster_id": "foo", "placement": "2", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:23:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:23:59 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd pool create", "pool": ".nfs"}]: dispatch 2026-03-10T10:23:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:59 vm01 ceph-mon[48597]: from='client.14512 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:23:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:59 vm01 ceph-mon[48597]: pgmap v91: 65 pgs: 52 active+clean, 10 creating+peering, 3 unknown; 449 KiB data, 49 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:23:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:59 vm01 ceph-mon[48597]: from='client.24321 -' entity='client.admin' cmd=[{"prefix": "nfs cluster create", "cluster_id": "foo", "placement": "2", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:23:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:23:59 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd pool create", "pool": ".nfs"}]: dispatch 2026-03-10T10:24:00.388 INFO:teuthology.orchestra.run.vm01.stdout:NFS Cluster Created Successfully 2026-03-10T10:24:00.436 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- bash -c 'ceph nfs export create cephfs --fsname foofs --clusterid foo --binding /fake || ceph nfs export create cephfs --fsname foofs --cluster-id foo --pseudo-path /fake' 2026-03-10T10:24:00.601 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:00 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "osd pool create", "pool": ".nfs"}]': finished 2026-03-10T10:24:00.601 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:00 vm01 ceph-mon[48597]: osdmap e50: 8 total, 8 up, 8 in 2026-03-10T10:24:00.601 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:00 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd pool application enable", "pool": ".nfs", "app": "nfs"}]: dispatch 2026-03-10T10:24:00.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:00 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "osd pool create", "pool": ".nfs"}]': finished 2026-03-10T10:24:00.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:00 vm08 ceph-mon[50506]: osdmap e50: 8 total, 8 up, 8 in 2026-03-10T10:24:00.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:00 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd pool application enable", "pool": ".nfs", "app": "nfs"}]: dispatch 2026-03-10T10:24:01.005 INFO:teuthology.orchestra.run.vm01.stderr:Invalid command: Unexpected argument '--clusterid' 2026-03-10T10:24:01.005 INFO:teuthology.orchestra.run.vm01.stderr:nfs export create cephfs [] [--readonly] [--client_addr ...] [--squash ] : Create a CephFS export 2026-03-10T10:24:01.005 INFO:teuthology.orchestra.run.vm01.stderr:Error EINVAL: invalid command 2026-03-10T10:24:01.315 INFO:teuthology.orchestra.run.vm01.stdout:{ 2026-03-10T10:24:01.315 INFO:teuthology.orchestra.run.vm01.stdout: "bind": "/fake", 2026-03-10T10:24:01.315 INFO:teuthology.orchestra.run.vm01.stdout: "fs": "foofs", 2026-03-10T10:24:01.315 INFO:teuthology.orchestra.run.vm01.stdout: "path": "/", 2026-03-10T10:24:01.315 INFO:teuthology.orchestra.run.vm01.stdout: "cluster": "foo", 2026-03-10T10:24:01.315 INFO:teuthology.orchestra.run.vm01.stdout: "mode": "RW" 2026-03-10T10:24:01.315 INFO:teuthology.orchestra.run.vm01.stdout:} 2026-03-10T10:24:01.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:01 vm01 ceph-mon[48597]: pgmap v93: 97 pgs: 9 creating+peering, 23 unknown, 65 active+clean; 449 KiB data, 50 MiB used, 160 GiB / 160 GiB avail; 519 B/s wr, 4 op/s 2026-03-10T10:24:01.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:01 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "osd pool application enable", "pool": ".nfs", "app": "nfs"}]': finished 2026-03-10T10:24:01.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:01 vm01 ceph-mon[48597]: osdmap e51: 8 total, 8 up, 8 in 2026-03-10T10:24:01.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:01 vm01 ceph-mon[48597]: Saving service nfs.foo spec with placement count:2 2026-03-10T10:24:01.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:01 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:24:01.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:01 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:24:01.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:01 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:24:01.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:01 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:24:01.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:01 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:24:01.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:01 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:24:01.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:01 vm01 ceph-mon[48597]: Creating key for client.nfs.foo.0.0.vm01.zvuakd 2026-03-10T10:24:01.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:01 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm01.zvuakd", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo"]}]: dispatch 2026-03-10T10:24:01.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:01 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm01.zvuakd", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo"]}]': finished 2026-03-10T10:24:01.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:01 vm01 ceph-mon[48597]: Ensuring nfs.foo.0 is in the ganesha grace table 2026-03-10T10:24:01.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:01 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.foo", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]: dispatch 2026-03-10T10:24:01.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:01 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.foo", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]': finished 2026-03-10T10:24:01.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:01 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:24:01.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:01 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.foo"}]: dispatch 2026-03-10T10:24:01.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:01 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.foo"}]': finished 2026-03-10T10:24:01.419 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:01 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm01.zvuakd-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-10T10:24:01.420 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:01 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm01.zvuakd-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-10T10:24:01.420 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:01 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:24:01.420 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:01 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo, allow rw tag cephfs data=foofs", "mds", "allow rw path=/"], "format": "json"}]: dispatch 2026-03-10T10:24:01.420 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:01 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo, allow rw tag cephfs data=foofs", "mds", "allow rw path=/"], "format": "json"}]': finished 2026-03-10T10:24:01.420 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6b3e19a-1c6a-11f1-aaba-ff800f171c2f -- bash -c 'while ! ceph orch ls | grep nfs | grep 2/2 ; do sleep 1 ; done' 2026-03-10T10:24:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:01 vm08 ceph-mon[50506]: pgmap v93: 97 pgs: 9 creating+peering, 23 unknown, 65 active+clean; 449 KiB data, 50 MiB used, 160 GiB / 160 GiB avail; 519 B/s wr, 4 op/s 2026-03-10T10:24:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:01 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "osd pool application enable", "pool": ".nfs", "app": "nfs"}]': finished 2026-03-10T10:24:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:01 vm08 ceph-mon[50506]: osdmap e51: 8 total, 8 up, 8 in 2026-03-10T10:24:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:01 vm08 ceph-mon[50506]: Saving service nfs.foo spec with placement count:2 2026-03-10T10:24:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:01 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:24:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:01 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:24:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:01 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:24:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:01 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:24:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:01 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:24:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:01 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:24:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:01 vm08 ceph-mon[50506]: Creating key for client.nfs.foo.0.0.vm01.zvuakd 2026-03-10T10:24:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:01 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm01.zvuakd", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo"]}]: dispatch 2026-03-10T10:24:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:01 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm01.zvuakd", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo"]}]': finished 2026-03-10T10:24:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:01 vm08 ceph-mon[50506]: Ensuring nfs.foo.0 is in the ganesha grace table 2026-03-10T10:24:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:01 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.foo", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]: dispatch 2026-03-10T10:24:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:01 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.foo", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]': finished 2026-03-10T10:24:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:01 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:24:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:01 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.foo"}]: dispatch 2026-03-10T10:24:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:01 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.foo"}]': finished 2026-03-10T10:24:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:01 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm01.zvuakd-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-10T10:24:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:01 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm01.zvuakd-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-10T10:24:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:01 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:24:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:01 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo, allow rw tag cephfs data=foofs", "mds", "allow rw path=/"], "format": "json"}]: dispatch 2026-03-10T10:24:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:01 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo, allow rw tag cephfs data=foofs", "mds", "allow rw path=/"], "format": "json"}]': finished 2026-03-10T10:24:02.483 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:02 vm08 ceph-mon[50506]: Rados config object exists: conf-nfs.foo 2026-03-10T10:24:02.483 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:02 vm08 ceph-mon[50506]: Creating key for client.nfs.foo.0.0.vm01.zvuakd-rgw 2026-03-10T10:24:02.483 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:02 vm08 ceph-mon[50506]: Deploying daemon nfs.foo.0.0.vm01.zvuakd on vm01 2026-03-10T10:24:02.483 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:02 vm08 ceph-mon[50506]: from='client.14536 -' entity='client.admin' cmd=[{"prefix": "nfs export create cephfs", "fsname": "foofs", "cluster_id": "foo", "pseudo_path": "/fake", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:24:02.483 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:02 vm08 ceph-mon[50506]: osdmap e52: 8 total, 8 up, 8 in 2026-03-10T10:24:02.483 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:02 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:24:02.483 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:02 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1.0.vm08.uzbdmx", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo"]}]: dispatch 2026-03-10T10:24:02.483 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:02 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1.0.vm08.uzbdmx", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo"]}]': finished 2026-03-10T10:24:02.483 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:02 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.foo", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]: dispatch 2026-03-10T10:24:02.483 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:02 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.foo", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]': finished 2026-03-10T10:24:02.483 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:02 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:24:02.483 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:02 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.foo"}]: dispatch 2026-03-10T10:24:02.483 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:02 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.foo"}]': finished 2026-03-10T10:24:02.483 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:02 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1.0.vm08.uzbdmx-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-10T10:24:02.483 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:02 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1.0.vm08.uzbdmx-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-10T10:24:02.483 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:02 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:24:02.692 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:02 vm01 ceph-mon[48597]: Rados config object exists: conf-nfs.foo 2026-03-10T10:24:02.692 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:02 vm01 ceph-mon[48597]: Creating key for client.nfs.foo.0.0.vm01.zvuakd-rgw 2026-03-10T10:24:02.692 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:02 vm01 ceph-mon[48597]: Deploying daemon nfs.foo.0.0.vm01.zvuakd on vm01 2026-03-10T10:24:02.692 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:02 vm01 ceph-mon[48597]: from='client.14536 -' entity='client.admin' cmd=[{"prefix": "nfs export create cephfs", "fsname": "foofs", "cluster_id": "foo", "pseudo_path": "/fake", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:24:02.692 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:02 vm01 ceph-mon[48597]: osdmap e52: 8 total, 8 up, 8 in 2026-03-10T10:24:02.692 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:02 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:24:02.692 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:02 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1.0.vm08.uzbdmx", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo"]}]: dispatch 2026-03-10T10:24:02.692 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:02 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1.0.vm08.uzbdmx", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo"]}]': finished 2026-03-10T10:24:02.692 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:02 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.foo", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]: dispatch 2026-03-10T10:24:02.692 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:02 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.foo", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]': finished 2026-03-10T10:24:02.692 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:02 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:24:02.692 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:02 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.foo"}]: dispatch 2026-03-10T10:24:02.692 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:02 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.foo"}]': finished 2026-03-10T10:24:02.692 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:02 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1.0.vm08.uzbdmx-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-10T10:24:02.692 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:02 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1.0.vm08.uzbdmx-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-10T10:24:02.692 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:02 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:24:03.495 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:03 vm08 ceph-mon[50506]: Creating key for client.nfs.foo.1.0.vm08.uzbdmx 2026-03-10T10:24:03.495 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:03 vm08 ceph-mon[50506]: Ensuring nfs.foo.1 is in the ganesha grace table 2026-03-10T10:24:03.495 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:03 vm08 ceph-mon[50506]: Rados config object exists: conf-nfs.foo 2026-03-10T10:24:03.495 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:03 vm08 ceph-mon[50506]: Creating key for client.nfs.foo.1.0.vm08.uzbdmx-rgw 2026-03-10T10:24:03.495 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:03 vm08 ceph-mon[50506]: Deploying daemon nfs.foo.1.0.vm08.uzbdmx on vm08 2026-03-10T10:24:03.495 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:03 vm08 ceph-mon[50506]: from='client.24359 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:24:03.495 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:03 vm08 ceph-mon[50506]: pgmap v96: 97 pgs: 9 creating+peering, 15 unknown, 73 active+clean; 451 KiB data, 50 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s wr, 7 op/s 2026-03-10T10:24:03.495 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:03 vm08 ceph-mon[50506]: mgrmap e20: vm01.nkkzmc(active, since 113s), standbys: vm08.rwpjaf 2026-03-10T10:24:03.495 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:03 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:24:03.495 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:03 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:24:03.495 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:03 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:24:03.495 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:03 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:24:03.499 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:03 vm01 ceph-mon[48597]: Creating key for client.nfs.foo.1.0.vm08.uzbdmx 2026-03-10T10:24:03.499 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:03 vm01 ceph-mon[48597]: Ensuring nfs.foo.1 is in the ganesha grace table 2026-03-10T10:24:03.499 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:03 vm01 ceph-mon[48597]: Rados config object exists: conf-nfs.foo 2026-03-10T10:24:03.499 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:03 vm01 ceph-mon[48597]: Creating key for client.nfs.foo.1.0.vm08.uzbdmx-rgw 2026-03-10T10:24:03.499 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:03 vm01 ceph-mon[48597]: Deploying daemon nfs.foo.1.0.vm08.uzbdmx on vm08 2026-03-10T10:24:03.499 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:03 vm01 ceph-mon[48597]: from='client.24359 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:24:03.500 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:03 vm01 ceph-mon[48597]: pgmap v96: 97 pgs: 9 creating+peering, 15 unknown, 73 active+clean; 451 KiB data, 50 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s wr, 7 op/s 2026-03-10T10:24:03.500 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:03 vm01 ceph-mon[48597]: mgrmap e20: vm01.nkkzmc(active, since 113s), standbys: vm08.rwpjaf 2026-03-10T10:24:03.500 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:03 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:24:03.500 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:03 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:24:03.500 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:03 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:24:03.500 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:03 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:24:04.395 INFO:teuthology.orchestra.run.vm01.stdout:nfs.foo ?:2049 2/2 0s ago 4s count:2 2026-03-10T10:24:04.436 INFO:teuthology.run_tasks:Running task vip.exec... 2026-03-10T10:24:04.438 INFO:tasks.vip:Running commands on role host.a host ubuntu@vm01.local 2026-03-10T10:24:04.438 DEBUG:teuthology.orchestra.run.vm01:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c 'mkdir /mnt/foo' 2026-03-10T10:24:04.461 INFO:teuthology.orchestra.run.vm01.stderr:+ mkdir /mnt/foo 2026-03-10T10:24:04.463 DEBUG:teuthology.orchestra.run.vm01:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c 'while ! mount -t nfs $(hostname):/fake /mnt/foo -o sync ; do sleep 5 ; done' 2026-03-10T10:24:04.526 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:24:04.526 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:24:04.693 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:04 vm01 ceph-mon[48597]: from='client.14562 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:24:04.693 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:04 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:24:04.693 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:04 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:24:04.693 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:04 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:24:04.693 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:04 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:24:04.693 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:04 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:24:04.693 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:04 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:24:04.693 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:04 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:24:04.693 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:04 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:24:04.693 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:04 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:24:04.734 INFO:teuthology.orchestra.run.vm01.stderr:Created symlink /run/systemd/system/remote-fs.target.wants/rpc-statd.service → /usr/lib/systemd/system/rpc-statd.service. 2026-03-10T10:24:04.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:04 vm08 ceph-mon[50506]: from='client.14562 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:24:04.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:04 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:24:04.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:04 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:24:04.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:04 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:24:04.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:04 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:24:04.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:04 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:24:04.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:04 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:24:04.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:04 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:24:04.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:04 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:24:04.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:04 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:24:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:05 vm08 ceph-mon[50506]: pgmap v97: 97 pgs: 9 creating+peering, 88 active+clean; 451 KiB data, 51 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 3.8 KiB/s wr, 12 op/s 2026-03-10T10:24:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:05 vm08 ceph-mon[50506]: from='client.14566 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:24:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:05 vm01 ceph-mon[48597]: pgmap v97: 97 pgs: 9 creating+peering, 88 active+clean; 451 KiB data, 51 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 3.8 KiB/s wr, 12 op/s 2026-03-10T10:24:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:05 vm01 ceph-mon[48597]: from='client.14566 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:24:07.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:07 vm08 ceph-mon[50506]: pgmap v98: 97 pgs: 97 active+clean; 452 KiB data, 51 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 3.9 KiB/s wr, 9 op/s 2026-03-10T10:24:07.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:07 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:24:07.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:07 vm01 ceph-mon[48597]: pgmap v98: 97 pgs: 97 active+clean; 452 KiB data, 51 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 3.9 KiB/s wr, 9 op/s 2026-03-10T10:24:07.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:07 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:24:08.314 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:24:08.314 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:24:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:09 vm08 ceph-mon[50506]: pgmap v99: 97 pgs: 97 active+clean; 452 KiB data, 51 MiB used, 160 GiB / 160 GiB avail; 3.7 KiB/s rd, 4.0 KiB/s wr, 9 op/s 2026-03-10T10:24:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:09 vm01 ceph-mon[48597]: pgmap v99: 97 pgs: 97 active+clean; 452 KiB data, 51 MiB used, 160 GiB / 160 GiB avail; 3.7 KiB/s rd, 4.0 KiB/s wr, 9 op/s 2026-03-10T10:24:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.f", "id": [7, 2]}]: dispatch 2026-03-10T10:24:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.7", "id": [1, 2]}]: dispatch 2026-03-10T10:24:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.f", "id": [1, 2]}]: dispatch 2026-03-10T10:24:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.1f", "id": [1, 2]}]: dispatch 2026-03-10T10:24:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:24:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:24:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:24:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.f", "id": [7, 2]}]: dispatch 2026-03-10T10:24:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.7", "id": [1, 2]}]: dispatch 2026-03-10T10:24:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.f", "id": [1, 2]}]: dispatch 2026-03-10T10:24:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.1f", "id": [1, 2]}]: dispatch 2026-03-10T10:24:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:24:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:24:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:24:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:11 vm08 ceph-mon[50506]: pgmap v100: 97 pgs: 97 active+clean; 452 KiB data, 51 MiB used, 160 GiB / 160 GiB avail; 3.1 KiB/s rd, 3.3 KiB/s wr, 7 op/s 2026-03-10T10:24:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:11 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.f", "id": [7, 2]}]': finished 2026-03-10T10:24:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:11 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.7", "id": [1, 2]}]': finished 2026-03-10T10:24:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:11 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.f", "id": [1, 2]}]': finished 2026-03-10T10:24:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:11 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.1f", "id": [1, 2]}]': finished 2026-03-10T10:24:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:11 vm08 ceph-mon[50506]: osdmap e53: 8 total, 8 up, 8 in 2026-03-10T10:24:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:11 vm01 ceph-mon[48597]: pgmap v100: 97 pgs: 97 active+clean; 452 KiB data, 51 MiB used, 160 GiB / 160 GiB avail; 3.1 KiB/s rd, 3.3 KiB/s wr, 7 op/s 2026-03-10T10:24:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.f", "id": [7, 2]}]': finished 2026-03-10T10:24:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.7", "id": [1, 2]}]': finished 2026-03-10T10:24:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.f", "id": [1, 2]}]': finished 2026-03-10T10:24:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.1f", "id": [1, 2]}]': finished 2026-03-10T10:24:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:11 vm01 ceph-mon[48597]: osdmap e53: 8 total, 8 up, 8 in 2026-03-10T10:24:12.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:12 vm08 ceph-mon[50506]: osdmap e54: 8 total, 8 up, 8 in 2026-03-10T10:24:12.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:12 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:24:12.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:12 vm01 ceph-mon[48597]: osdmap e54: 8 total, 8 up, 8 in 2026-03-10T10:24:12.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:12 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:24:13.315 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:24:13.316 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:24:13.342 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:24:13.342 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:24:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:13 vm08 ceph-mon[50506]: pgmap v103: 97 pgs: 1 peering, 96 active+clean; 452 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 1.9 KiB/s rd, 1.6 KiB/s wr, 3 op/s 2026-03-10T10:24:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:13 vm01 ceph-mon[48597]: pgmap v103: 97 pgs: 1 peering, 96 active+clean; 452 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 1.9 KiB/s rd, 1.6 KiB/s wr, 3 op/s 2026-03-10T10:24:15.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:15 vm08 ceph-mon[50506]: pgmap v104: 97 pgs: 1 peering, 96 active+clean; 452 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 2.0 KiB/s rd, 1.1 KiB/s wr, 3 op/s 2026-03-10T10:24:15.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:15 vm01 ceph-mon[48597]: pgmap v104: 97 pgs: 1 peering, 96 active+clean; 452 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 2.0 KiB/s rd, 1.1 KiB/s wr, 3 op/s 2026-03-10T10:24:17.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:17 vm01 ceph-mon[48597]: pgmap v105: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 383 B/s rd, 895 B/s wr, 1 op/s; 2 B/s, 0 objects/s recovering 2026-03-10T10:24:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:17 vm08 ceph-mon[50506]: pgmap v105: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 383 B/s rd, 895 B/s wr, 1 op/s; 2 B/s, 0 objects/s recovering 2026-03-10T10:24:18.343 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:24:18.344 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:24:18.369 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:24:18.369 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:24:19.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:19 vm08 ceph-mon[50506]: pgmap v106: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 383 B/s rd, 895 B/s wr, 1 op/s; 2 B/s, 0 objects/s recovering 2026-03-10T10:24:19.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:19 vm01 ceph-mon[48597]: pgmap v106: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 383 B/s rd, 895 B/s wr, 1 op/s; 2 B/s, 0 objects/s recovering 2026-03-10T10:24:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:21 vm01 ceph-mon[48597]: pgmap v107: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 320 B/s rd, 746 B/s wr, 1 op/s; 2 B/s, 0 objects/s recovering 2026-03-10T10:24:21.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:21 vm08 ceph-mon[50506]: pgmap v107: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 320 B/s rd, 746 B/s wr, 1 op/s; 2 B/s, 0 objects/s recovering 2026-03-10T10:24:23.370 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:24:23.371 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:24:23.396 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:24:23.396 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:24:23.513 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:23 vm01 ceph-mon[48597]: pgmap v108: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 289 B/s rd, 676 B/s wr, 1 op/s; 2 B/s, 0 objects/s recovering 2026-03-10T10:24:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:23 vm08 ceph-mon[50506]: pgmap v108: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 289 B/s rd, 676 B/s wr, 1 op/s; 2 B/s, 0 objects/s recovering 2026-03-10T10:24:25.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:25 vm08 ceph-mon[50506]: pgmap v109: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 426 B/s rd, 597 B/s wr, 1 op/s; 1 B/s, 0 objects/s recovering 2026-03-10T10:24:25.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:25 vm01 ceph-mon[48597]: pgmap v109: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 426 B/s rd, 597 B/s wr, 1 op/s; 1 B/s, 0 objects/s recovering 2026-03-10T10:24:27.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:27 vm08 ceph-mon[50506]: pgmap v110: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 597 B/s wr, 0 op/s; 1 B/s, 0 objects/s recovering 2026-03-10T10:24:27.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:27 vm01 ceph-mon[48597]: pgmap v110: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 597 B/s wr, 0 op/s; 1 B/s, 0 objects/s recovering 2026-03-10T10:24:28.397 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:24:28.398 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:24:28.424 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:24:28.425 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:24:29.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:29 vm08 ceph-mon[50506]: pgmap v111: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:24:29.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:29 vm01 ceph-mon[48597]: pgmap v111: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:24:31.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:31 vm08 ceph-mon[50506]: pgmap v112: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:24:31.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:31 vm01 ceph-mon[48597]: pgmap v112: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:24:33.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:33 vm01 ceph-mon[48597]: pgmap v113: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:24:33.426 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:24:33.427 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:24:33.455 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:24:33.455 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:24:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:33 vm08 ceph-mon[50506]: pgmap v113: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:24:35.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:35 vm08 ceph-mon[50506]: pgmap v114: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:24:35.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:35 vm01 ceph-mon[48597]: pgmap v114: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:24:37.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:37 vm08 ceph-mon[50506]: pgmap v115: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:24:37.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:37 vm01 ceph-mon[48597]: pgmap v115: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:24:38.457 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:24:38.457 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:24:38.507 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:24:38.508 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:24:39.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:39 vm01 ceph-mon[48597]: pgmap v116: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:24:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:39 vm08 ceph-mon[50506]: pgmap v116: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:24:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:41 vm01 ceph-mon[48597]: pgmap v117: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:24:41.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:41 vm08 ceph-mon[50506]: pgmap v117: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:24:43.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:43 vm01 ceph-mon[48597]: pgmap v118: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:24:43.509 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:24:43.510 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:24:43.534 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:24:43.535 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:24:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:43 vm08 ceph-mon[50506]: pgmap v118: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:24:45.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:45 vm01 ceph-mon[48597]: pgmap v119: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:24:45.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:45 vm08 ceph-mon[50506]: pgmap v119: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:24:47.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:47 vm01 ceph-mon[48597]: pgmap v120: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:24:47.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:47 vm08 ceph-mon[50506]: pgmap v120: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:24:48.536 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:24:48.536 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:24:48.592 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:24:48.592 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:24:49.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:49 vm08 ceph-mon[50506]: pgmap v121: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:24:49.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:49 vm01 ceph-mon[48597]: pgmap v121: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:24:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:51 vm01 ceph-mon[48597]: pgmap v122: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:24:51.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:51 vm08 ceph-mon[50506]: pgmap v122: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:24:53.515 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:53 vm01 ceph-mon[48597]: pgmap v123: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:24:53.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:53 vm08 ceph-mon[50506]: pgmap v123: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:24:53.594 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:24:53.594 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:24:53.621 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:24:53.621 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:24:55.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:55 vm08 ceph-mon[50506]: pgmap v124: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:24:55.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:55 vm01 ceph-mon[48597]: pgmap v124: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:24:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:57 vm08 ceph-mon[50506]: pgmap v125: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:24:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:57 vm01 ceph-mon[48597]: pgmap v125: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:24:58.623 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:24:58.623 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:24:58.649 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:24:58.650 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:24:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:24:59 vm08 ceph-mon[50506]: pgmap v126: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:24:59.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:24:59 vm01 ceph-mon[48597]: pgmap v126: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:01.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:01 vm08 ceph-mon[50506]: pgmap v127: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:01.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:01 vm01 ceph-mon[48597]: pgmap v127: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:03 vm08 ceph-mon[50506]: pgmap v128: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:25:03.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:03 vm01 ceph-mon[48597]: pgmap v128: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:25:03.651 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:25:03.651 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:25:03.677 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:25:03.678 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:25:04.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:04 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:25:04.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:04 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:25:04.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:04 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:25:04.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:04 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:25:04.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:04 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:25:04.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:04 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:25:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:05 vm08 ceph-mon[50506]: pgmap v129: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:05 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:25:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:05 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:25:05.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:05 vm01 ceph-mon[48597]: pgmap v129: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:05.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:05 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:25:05.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:05 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:25:06.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:06 vm08 ceph-mon[50506]: pgmap v130: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:25:06.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:06 vm01 ceph-mon[48597]: pgmap v130: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:25:08.679 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:25:08.680 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:25:08.707 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:25:08.708 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:25:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:09 vm08 ceph-mon[50506]: pgmap v131: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:09.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:09 vm01 ceph-mon[48597]: pgmap v131: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:10.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:25:10.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:25:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:25:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:25:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:11 vm08 ceph-mon[50506]: pgmap v132: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:11.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:11 vm01 ceph-mon[48597]: pgmap v132: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:13.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:13 vm08 ceph-mon[50506]: pgmap v133: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:25:13.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:13 vm01 ceph-mon[48597]: pgmap v133: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:25:13.709 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:25:13.710 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:25:13.736 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:25:13.736 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:25:15.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:15 vm08 ceph-mon[50506]: pgmap v134: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:15.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:15 vm01 ceph-mon[48597]: pgmap v134: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:17 vm08 ceph-mon[50506]: pgmap v135: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:25:17.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:17 vm01 ceph-mon[48597]: pgmap v135: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:25:18.737 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:25:18.738 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:25:18.764 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:25:18.764 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:25:19.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:19 vm08 ceph-mon[50506]: pgmap v136: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:19.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:19 vm01 ceph-mon[48597]: pgmap v136: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:21.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:21 vm08 ceph-mon[50506]: pgmap v137: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:21.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:21 vm01 ceph-mon[48597]: pgmap v137: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:23.766 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:25:23.766 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:25:23.792 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:25:23.793 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:25:23.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:23 vm08 ceph-mon[50506]: pgmap v138: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:25:23.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:23 vm01 ceph-mon[48597]: pgmap v138: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:25:25.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:25 vm08 ceph-mon[50506]: pgmap v139: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:25 vm01 ceph-mon[48597]: pgmap v139: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:27.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:27 vm08 ceph-mon[50506]: pgmap v140: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:25:27.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:27 vm01 ceph-mon[48597]: pgmap v140: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:25:28.794 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:25:28.795 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:25:28.821 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:25:28.821 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:25:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:29 vm08 ceph-mon[50506]: pgmap v141: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:29.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:29 vm01 ceph-mon[48597]: pgmap v141: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:31.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:31 vm08 ceph-mon[50506]: pgmap v142: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:31.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:31 vm01 ceph-mon[48597]: pgmap v142: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:33.823 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:25:33.823 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:25:33.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:33 vm08 ceph-mon[50506]: pgmap v143: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:25:33.849 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:25:33.849 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:25:33.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:33 vm01 ceph-mon[48597]: pgmap v143: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:25:35.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:35 vm08 ceph-mon[50506]: pgmap v144: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:35.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:35 vm01 ceph-mon[48597]: pgmap v144: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:37.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:37 vm08 ceph-mon[50506]: pgmap v145: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:25:37.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:37 vm01 ceph-mon[48597]: pgmap v145: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:25:38.851 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:25:38.851 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:25:38.876 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:25:38.877 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:25:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:39 vm08 ceph-mon[50506]: pgmap v146: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:39.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:39 vm01 ceph-mon[48597]: pgmap v146: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:41.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:41 vm08 ceph-mon[50506]: pgmap v147: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:41.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:41 vm01 ceph-mon[48597]: pgmap v147: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:43.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:43 vm08 ceph-mon[50506]: pgmap v148: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:25:43.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:43 vm01 ceph-mon[48597]: pgmap v148: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:25:43.878 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:25:43.879 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:25:43.904 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:25:43.905 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:25:45.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:45 vm08 ceph-mon[50506]: pgmap v149: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:45.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:45 vm01 ceph-mon[48597]: pgmap v149: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:47.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:47 vm08 ceph-mon[50506]: pgmap v150: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:25:47.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:47 vm01 ceph-mon[48597]: pgmap v150: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:25:48.906 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:25:48.907 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:25:48.933 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:25:48.933 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:25:49.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:49 vm08 ceph-mon[50506]: pgmap v151: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:49.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:49 vm01 ceph-mon[48597]: pgmap v151: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:51.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:51 vm08 ceph-mon[50506]: pgmap v152: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:51.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:51 vm01 ceph-mon[48597]: pgmap v152: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:52.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:52 vm08 ceph-mon[50506]: pgmap v153: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:25:52.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:52 vm01 ceph-mon[48597]: pgmap v153: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:25:53.934 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:25:53.935 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:25:53.962 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:25:53.962 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:25:55.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:55 vm08 ceph-mon[50506]: pgmap v154: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:55.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:55 vm01 ceph-mon[48597]: pgmap v154: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:57 vm08 ceph-mon[50506]: pgmap v155: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:25:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:57 vm01 ceph-mon[48597]: pgmap v155: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:25:58.964 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:25:58.964 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:25:58.990 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:25:58.990 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:25:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:25:59 vm08 ceph-mon[50506]: pgmap v156: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:25:59.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:25:59 vm01 ceph-mon[48597]: pgmap v156: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:01.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:01 vm08 ceph-mon[50506]: pgmap v157: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:01.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:01 vm01 ceph-mon[48597]: pgmap v157: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:03.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:03 vm08 ceph-mon[50506]: pgmap v158: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:26:03.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:03 vm01 ceph-mon[48597]: pgmap v158: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:26:03.991 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:26:03.992 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:26:04.019 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:26:04.020 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:26:05.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:05 vm08 ceph-mon[50506]: pgmap v159: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:05.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:05 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:26:05.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:05 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:26:05.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:05 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:26:05.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:05 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:26:05.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:05 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:26:05.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:05 vm01 ceph-mon[48597]: pgmap v159: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:05.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:05 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:26:05.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:05 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:26:05.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:05 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:26:05.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:05 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:26:05.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:05 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:26:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:07 vm08 ceph-mon[50506]: pgmap v160: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:26:07.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:07 vm01 ceph-mon[48597]: pgmap v160: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:26:09.021 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:26:09.022 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:26:09.047 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:26:09.048 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:26:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:09 vm08 ceph-mon[50506]: pgmap v161: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:09.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:09 vm01 ceph-mon[48597]: pgmap v161: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:10.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:26:10.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:26:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:26:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:26:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:11 vm08 ceph-mon[50506]: pgmap v162: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:11.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:11 vm01 ceph-mon[48597]: pgmap v162: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:13.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:13 vm08 ceph-mon[50506]: pgmap v163: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:26:13.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:13 vm01 ceph-mon[48597]: pgmap v163: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:26:14.049 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:26:14.050 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:26:14.080 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:26:14.080 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:26:15.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:15 vm08 ceph-mon[50506]: pgmap v164: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:15.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:15 vm01 ceph-mon[48597]: pgmap v164: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:17 vm08 ceph-mon[50506]: pgmap v165: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:26:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:17 vm01 ceph-mon[48597]: pgmap v165: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:26:19.082 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:26:19.082 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:26:19.109 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:26:19.110 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:26:19.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:19 vm08 ceph-mon[50506]: pgmap v166: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:19.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:19 vm01 ceph-mon[48597]: pgmap v166: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:21.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:21 vm08 ceph-mon[50506]: pgmap v167: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:21.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:21 vm01 ceph-mon[48597]: pgmap v167: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:23.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:23 vm08 ceph-mon[50506]: pgmap v168: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:26:23.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:23 vm01 ceph-mon[48597]: pgmap v168: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:26:24.111 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:26:24.111 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:26:24.136 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:26:24.137 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:26:25.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:25 vm01 ceph-mon[48597]: pgmap v169: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:25.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:25 vm08 ceph-mon[50506]: pgmap v169: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:27 vm01 ceph-mon[48597]: pgmap v170: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:26:27.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:27 vm08 ceph-mon[50506]: pgmap v170: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:26:29.138 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:26:29.139 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:26:29.171 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:26:29.171 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:26:29.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:29 vm01 ceph-mon[48597]: pgmap v171: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:29.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:29 vm08 ceph-mon[50506]: pgmap v171: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:31.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:31 vm01 ceph-mon[48597]: pgmap v172: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:31.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:31 vm08 ceph-mon[50506]: pgmap v172: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:33.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:33 vm08 ceph-mon[50506]: pgmap v173: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:26:33.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:33 vm01 ceph-mon[48597]: pgmap v173: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:26:34.173 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:26:34.173 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:26:34.244 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:26:34.244 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:26:35.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:35 vm08 ceph-mon[50506]: pgmap v174: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:35.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:35 vm01 ceph-mon[48597]: pgmap v174: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:37.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:37 vm08 ceph-mon[50506]: pgmap v175: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:26:37.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:37 vm01 ceph-mon[48597]: pgmap v175: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:26:39.246 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:26:39.246 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:26:39.271 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:26:39.272 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:26:39.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:39 vm08 ceph-mon[50506]: pgmap v176: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:39.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:39 vm01 ceph-mon[48597]: pgmap v176: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:41.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:41 vm08 ceph-mon[50506]: pgmap v177: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:41.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:41 vm01 ceph-mon[48597]: pgmap v177: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:43.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:43 vm08 ceph-mon[50506]: pgmap v178: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:26:43.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:43 vm01 ceph-mon[48597]: pgmap v178: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:26:44.273 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:26:44.274 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:26:44.344 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:26:44.344 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:26:45.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:45 vm08 ceph-mon[50506]: pgmap v179: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:45.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:45 vm01 ceph-mon[48597]: pgmap v179: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:47.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:47 vm08 ceph-mon[50506]: pgmap v180: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:26:47.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:47 vm01 ceph-mon[48597]: pgmap v180: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:26:49.346 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:26:49.346 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:26:49.374 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:26:49.375 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:26:49.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:49 vm08 ceph-mon[50506]: pgmap v181: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:49.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:49 vm01 ceph-mon[48597]: pgmap v181: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:51.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:51 vm08 ceph-mon[50506]: pgmap v182: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:51.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:51 vm01 ceph-mon[48597]: pgmap v182: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:53.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:53 vm08 ceph-mon[50506]: pgmap v183: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:26:53.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:53 vm01 ceph-mon[48597]: pgmap v183: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:26:54.376 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:26:54.378 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:26:54.403 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:26:54.404 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:26:55.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:55 vm08 ceph-mon[50506]: pgmap v184: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:55.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:55 vm01 ceph-mon[48597]: pgmap v184: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:57.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:57 vm08 ceph-mon[50506]: pgmap v185: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:26:57.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:57 vm01 ceph-mon[48597]: pgmap v185: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:26:59.405 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:26:59.406 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:26:59.432 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:26:59.432 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:26:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:26:59 vm08 ceph-mon[50506]: pgmap v186: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:26:59.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:26:59 vm01 ceph-mon[48597]: pgmap v186: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:01 vm08 ceph-mon[50506]: pgmap v187: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:01.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:01 vm01 ceph-mon[48597]: pgmap v187: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:02.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:02 vm08 ceph-mon[50506]: pgmap v188: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:27:02.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:02 vm01 ceph-mon[48597]: pgmap v188: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:27:04.434 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:27:04.435 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:27:04.462 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:27:04.462 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:27:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:05 vm08 ceph-mon[50506]: pgmap v189: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:05 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:27:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:05 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:27:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:05 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:27:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:05 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:27:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:05 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:27:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:05 vm01 ceph-mon[48597]: pgmap v189: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:05 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:27:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:05 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:27:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:05 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:27:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:05 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:27:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:05 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:27:07.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:07 vm08 ceph-mon[50506]: pgmap v190: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:27:07.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:07 vm01 ceph-mon[48597]: pgmap v190: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:27:09.464 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:27:09.464 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:27:09.490 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:27:09.490 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:27:09.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:09 vm08 ceph-mon[50506]: pgmap v191: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:09.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:09 vm01 ceph-mon[48597]: pgmap v191: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:27:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:27:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:27:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:27:11.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:11 vm08 ceph-mon[50506]: pgmap v192: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:11 vm01 ceph-mon[48597]: pgmap v192: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:13.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:13 vm08 ceph-mon[50506]: pgmap v193: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:27:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:13 vm01 ceph-mon[48597]: pgmap v193: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:27:14.492 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:27:14.492 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:27:14.521 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:27:14.521 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:27:15.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:15 vm08 ceph-mon[50506]: pgmap v194: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:15.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:15 vm01 ceph-mon[48597]: pgmap v194: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:17.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:17 vm08 ceph-mon[50506]: pgmap v195: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:27:17.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:17 vm01 ceph-mon[48597]: pgmap v195: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:27:19.523 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:27:19.523 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:27:19.552 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:27:19.552 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:27:19.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:19 vm08 ceph-mon[50506]: pgmap v196: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:19.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:19 vm01 ceph-mon[48597]: pgmap v196: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:21.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:21 vm08 ceph-mon[50506]: pgmap v197: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:21 vm01 ceph-mon[48597]: pgmap v197: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:23.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:23 vm08 ceph-mon[50506]: pgmap v198: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:27:23.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:23 vm01 ceph-mon[48597]: pgmap v198: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:27:24.554 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:27:24.554 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:27:24.580 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:27:24.581 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:27:25.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:25 vm08 ceph-mon[50506]: pgmap v199: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:25 vm01 ceph-mon[48597]: pgmap v199: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:27.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:27 vm08 ceph-mon[50506]: pgmap v200: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:27:27.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:27 vm01 ceph-mon[48597]: pgmap v200: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:27:29.582 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:27:29.583 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:27:29.610 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:27:29.610 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:27:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:29 vm08 ceph-mon[50506]: pgmap v201: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:29 vm01 ceph-mon[48597]: pgmap v201: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:31.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:31 vm08 ceph-mon[50506]: pgmap v202: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:31.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:31 vm01 ceph-mon[48597]: pgmap v202: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:33.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:33 vm08 ceph-mon[50506]: pgmap v203: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:27:33.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:33 vm01 ceph-mon[48597]: pgmap v203: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:27:34.612 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:27:34.612 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:27:34.639 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:27:34.639 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:27:35.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:35 vm08 ceph-mon[50506]: pgmap v204: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:35.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:35 vm01 ceph-mon[48597]: pgmap v204: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:36.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:36 vm08 ceph-mon[50506]: pgmap v205: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:27:36.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:36 vm01 ceph-mon[48597]: pgmap v205: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:27:39.547 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:39 vm08 ceph-mon[50506]: pgmap v206: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:39.570 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:39 vm01 ceph-mon[48597]: pgmap v206: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:39.640 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:27:39.641 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:27:39.672 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:27:39.673 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:27:41.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:41 vm08 ceph-mon[50506]: pgmap v207: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:41.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:41 vm01 ceph-mon[48597]: pgmap v207: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:43.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:43 vm08 ceph-mon[50506]: pgmap v208: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:27:43.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:43 vm01 ceph-mon[48597]: pgmap v208: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:27:44.674 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:27:44.675 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:27:44.706 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:27:44.707 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:27:45.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:45 vm08 ceph-mon[50506]: pgmap v209: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:45.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:45 vm01 ceph-mon[48597]: pgmap v209: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:47.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:47 vm08 ceph-mon[50506]: pgmap v210: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:27:47.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:47 vm01 ceph-mon[48597]: pgmap v210: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:27:49.708 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:27:49.709 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:27:49.734 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:27:49.735 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:27:49.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:49 vm08 ceph-mon[50506]: pgmap v211: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:49.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:49 vm01 ceph-mon[48597]: pgmap v211: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:51.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:51 vm08 ceph-mon[50506]: pgmap v212: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:51.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:51 vm01 ceph-mon[48597]: pgmap v212: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:53.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:53 vm08 ceph-mon[50506]: pgmap v213: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:27:53.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:53 vm01 ceph-mon[48597]: pgmap v213: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:27:54.736 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:27:54.736 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:27:54.763 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:27:54.764 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:27:55.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:55 vm08 ceph-mon[50506]: pgmap v214: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:55.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:55 vm01 ceph-mon[48597]: pgmap v214: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:57.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:57 vm08 ceph-mon[50506]: pgmap v215: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:27:57.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:57 vm01 ceph-mon[48597]: pgmap v215: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:27:59.766 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:27:59.766 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:27:59.793 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:27:59.794 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:27:59.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:27:59 vm08 ceph-mon[50506]: pgmap v216: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:27:59.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:27:59 vm01 ceph-mon[48597]: pgmap v216: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:01 vm08 ceph-mon[50506]: pgmap v217: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:01.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:01 vm01 ceph-mon[48597]: pgmap v217: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:03.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:03 vm08 ceph-mon[50506]: pgmap v218: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:28:03.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:03 vm01 ceph-mon[48597]: pgmap v218: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:28:04.795 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:28:04.796 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:28:04.823 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:28:04.823 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:28:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:05 vm08 ceph-mon[50506]: pgmap v219: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:05 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:28:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:05 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:28:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:05 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:28:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:05 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:28:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:05 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:28:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:05 vm01 ceph-mon[48597]: pgmap v219: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:05 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:28:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:05 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:28:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:05 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:28:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:05 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:28:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:05 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:28:07.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:07 vm08 ceph-mon[50506]: pgmap v220: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:28:07.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:07 vm01 ceph-mon[48597]: pgmap v220: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:28:09.825 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:28:09.825 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:28:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:09 vm08 ceph-mon[50506]: pgmap v221: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:09.851 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:28:09.851 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:28:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:09 vm01 ceph-mon[48597]: pgmap v221: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:28:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:28:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:28:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:28:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:11 vm08 ceph-mon[50506]: pgmap v222: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:11 vm01 ceph-mon[48597]: pgmap v222: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:12.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:12 vm08 ceph-mon[50506]: pgmap v223: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:28:12.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:12 vm01 ceph-mon[48597]: pgmap v223: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:28:14.853 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:28:14.853 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:28:14.906 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:28:14.906 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:28:15.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:15 vm08 ceph-mon[50506]: pgmap v224: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:15.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:15 vm01 ceph-mon[48597]: pgmap v224: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:17.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:17 vm08 ceph-mon[50506]: pgmap v225: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:28:17.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:17 vm01 ceph-mon[48597]: pgmap v225: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:28:18.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:18 vm08 ceph-mon[50506]: pgmap v226: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:18.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:18 vm01 ceph-mon[48597]: pgmap v226: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:19.908 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:28:19.908 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:28:19.934 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:28:19.934 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:28:21.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:21 vm08 ceph-mon[50506]: pgmap v227: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:21 vm01 ceph-mon[48597]: pgmap v227: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:23.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:23 vm08 ceph-mon[50506]: pgmap v228: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:28:23.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:23 vm01 ceph-mon[48597]: pgmap v228: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:28:24.936 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:28:24.936 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:28:24.961 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:28:24.961 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:28:25.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:25 vm08 ceph-mon[50506]: pgmap v229: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:25.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:25 vm01 ceph-mon[48597]: pgmap v229: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:27.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:27 vm08 ceph-mon[50506]: pgmap v230: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:28:27.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:27 vm01 ceph-mon[48597]: pgmap v230: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:28:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:29 vm08 ceph-mon[50506]: pgmap v231: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:29 vm01 ceph-mon[48597]: pgmap v231: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:29.962 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:28:29.963 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:28:29.990 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:28:29.990 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:28:31.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:31 vm08 ceph-mon[50506]: pgmap v232: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:31.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:31 vm01 ceph-mon[48597]: pgmap v232: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:33.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:33 vm08 ceph-mon[50506]: pgmap v233: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:28:33.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:33 vm01 ceph-mon[48597]: pgmap v233: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:28:34.991 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:28:34.992 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:28:35.021 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:28:35.021 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:28:35.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:35 vm08 ceph-mon[50506]: pgmap v234: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:35.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:35 vm01 ceph-mon[48597]: pgmap v234: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:37.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:37 vm08 ceph-mon[50506]: pgmap v235: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:28:37.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:37 vm01 ceph-mon[48597]: pgmap v235: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:28:38.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:38 vm08 ceph-mon[50506]: pgmap v236: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:38.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:38 vm01 ceph-mon[48597]: pgmap v236: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:40.023 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:28:40.023 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:28:40.051 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:28:40.052 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:28:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:40 vm08 ceph-mon[50506]: pgmap v237: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:40.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:40 vm01 ceph-mon[48597]: pgmap v237: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:43.532 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:43 vm01 ceph-mon[48597]: pgmap v238: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:28:43.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:43 vm08 ceph-mon[50506]: pgmap v238: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:28:45.054 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:28:45.054 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:28:45.079 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:28:45.080 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:28:45.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:45 vm08 ceph-mon[50506]: pgmap v239: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:45.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:45 vm01 ceph-mon[48597]: pgmap v239: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:47.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:47 vm08 ceph-mon[50506]: pgmap v240: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:28:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:47 vm01 ceph-mon[48597]: pgmap v240: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:28:49.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:49 vm08 ceph-mon[50506]: pgmap v241: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:49.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:49 vm01 ceph-mon[48597]: pgmap v241: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:50.081 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:28:50.082 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:28:50.107 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:28:50.108 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:28:50.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:50 vm08 ceph-mon[50506]: pgmap v242: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:50.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:50 vm01 ceph-mon[48597]: pgmap v242: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:53.533 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:53 vm01 ceph-mon[48597]: pgmap v243: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:28:53.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:53 vm08 ceph-mon[50506]: pgmap v243: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:28:55.109 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:28:55.110 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:28:55.135 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:28:55.136 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:28:55.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:55 vm08 ceph-mon[50506]: pgmap v244: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:55 vm01 ceph-mon[48597]: pgmap v244: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:57.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:57 vm08 ceph-mon[50506]: pgmap v245: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:28:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:57 vm01 ceph-mon[48597]: pgmap v245: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:28:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:28:59 vm08 ceph-mon[50506]: pgmap v246: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:28:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:28:59 vm01 ceph-mon[48597]: pgmap v246: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:00.137 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:29:00.138 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:29:00.163 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:29:00.164 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:29:00.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:00 vm08 ceph-mon[50506]: pgmap v247: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:00.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:00 vm01 ceph-mon[48597]: pgmap v247: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:03.534 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:03 vm01 ceph-mon[48597]: pgmap v248: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:29:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:03 vm08 ceph-mon[50506]: pgmap v248: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:29:05.165 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:29:05.166 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:29:05.192 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:29:05.193 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:29:05.497 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:05 vm01 ceph-mon[48597]: pgmap v249: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:05.508 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:05 vm08 ceph-mon[50506]: pgmap v249: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:06.478 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:06 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:29:06.478 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:06 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:29:06.478 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:06 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:29:06.478 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:06 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:29:06.478 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:06 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:29:06.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:06 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:29:06.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:06 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:29:06.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:06 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:29:06.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:06 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:29:06.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:06 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:29:07.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:07 vm08 ceph-mon[50506]: pgmap v250: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:29:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:07 vm01 ceph-mon[48597]: pgmap v250: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:29:09.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:09 vm08 ceph-mon[50506]: pgmap v251: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:09.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:09 vm01 ceph-mon[48597]: pgmap v251: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:10.194 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:29:10.195 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:29:10.220 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:29:10.220 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:29:10.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:29:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:29:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:10 vm08 ceph-mon[50506]: pgmap v252: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:29:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:29:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:10 vm01 ceph-mon[48597]: pgmap v252: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:13.535 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:13 vm01 ceph-mon[48597]: pgmap v253: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:29:13.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:13 vm08 ceph-mon[50506]: pgmap v253: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:29:15.222 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:29:15.222 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:29:15.248 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:29:15.248 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:29:15.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:15 vm08 ceph-mon[50506]: pgmap v254: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:15.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:15 vm01 ceph-mon[48597]: pgmap v254: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:17.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:17 vm08 ceph-mon[50506]: pgmap v255: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:29:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:17 vm01 ceph-mon[48597]: pgmap v255: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:29:19.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:19 vm08 ceph-mon[50506]: pgmap v256: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:19 vm01 ceph-mon[48597]: pgmap v256: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:20.250 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:29:20.251 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:29:20.308 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:29:20.308 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:29:20.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:20 vm08 ceph-mon[50506]: pgmap v257: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:20.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:20 vm01 ceph-mon[48597]: pgmap v257: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:23.535 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:23 vm01 ceph-mon[48597]: pgmap v258: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:29:23.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:23 vm08 ceph-mon[50506]: pgmap v258: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:29:25.309 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:29:25.310 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:29:25.335 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:29:25.336 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:29:25.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:25 vm08 ceph-mon[50506]: pgmap v259: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:25.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:25 vm01 ceph-mon[48597]: pgmap v259: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:27.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:27 vm08 ceph-mon[50506]: pgmap v260: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:29:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:27 vm01 ceph-mon[48597]: pgmap v260: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:29:29.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:29 vm08 ceph-mon[50506]: pgmap v261: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:29.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:29 vm01 ceph-mon[48597]: pgmap v261: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:30.337 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:29:30.337 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:29:30.363 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:29:30.363 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:29:30.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:30 vm08 ceph-mon[50506]: pgmap v262: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:30 vm01 ceph-mon[48597]: pgmap v262: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:33.536 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:33 vm01 ceph-mon[48597]: pgmap v263: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:29:33.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:33 vm08 ceph-mon[50506]: pgmap v263: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:29:35.365 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:29:35.365 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:29:35.392 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:29:35.392 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:29:35.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:35 vm08 ceph-mon[50506]: pgmap v264: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:35.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:35 vm01 ceph-mon[48597]: pgmap v264: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:37.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:37 vm08 ceph-mon[50506]: pgmap v265: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:29:37.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:37 vm01 ceph-mon[48597]: pgmap v265: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:29:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:39 vm08 ceph-mon[50506]: pgmap v266: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:39.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:39 vm01 ceph-mon[48597]: pgmap v266: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:40.395 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:29:40.395 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:29:40.423 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:29:40.424 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:29:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:40 vm08 ceph-mon[50506]: pgmap v267: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:40.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:40 vm01 ceph-mon[48597]: pgmap v267: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:43.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:42 vm08 ceph-mon[50506]: pgmap v268: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:29:43.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:42 vm01 ceph-mon[48597]: pgmap v268: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:29:45.426 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:29:45.427 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:29:45.511 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:29:45.512 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:29:45.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:45 vm08 ceph-mon[50506]: pgmap v269: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:45.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:45 vm01 ceph-mon[48597]: pgmap v269: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:47.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:47 vm08 ceph-mon[50506]: pgmap v270: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:29:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:47 vm01 ceph-mon[48597]: pgmap v270: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:29:49.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:49 vm08 ceph-mon[50506]: pgmap v271: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:49.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:49 vm01 ceph-mon[48597]: pgmap v271: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:50.513 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:29:50.514 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:29:50.544 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:29:50.545 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:29:50.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:50 vm08 ceph-mon[50506]: pgmap v272: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:50.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:50 vm01 ceph-mon[48597]: pgmap v272: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:53.538 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:53 vm01 ceph-mon[48597]: pgmap v273: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:53.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:53 vm08 ceph-mon[50506]: pgmap v273: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:55.546 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:29:55.547 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:29:55.572 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:29:55.573 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:29:55.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:55 vm08 ceph-mon[50506]: pgmap v274: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T10:29:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:55 vm01 ceph-mon[48597]: pgmap v274: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T10:29:57.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:57 vm08 ceph-mon[50506]: pgmap v275: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:57 vm01 ceph-mon[48597]: pgmap v275: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:29:59.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:29:59 vm08 ceph-mon[50506]: pgmap v276: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T10:29:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:29:59 vm01 ceph-mon[48597]: pgmap v276: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T10:30:00.574 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:30:00.575 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:30:00.600 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:30:00.600 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:30:00.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:00 vm08 ceph-mon[50506]: overall HEALTH_OK 2026-03-10T10:30:00.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:00 vm08 ceph-mon[50506]: pgmap v277: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T10:30:00.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:00 vm01 ceph-mon[48597]: overall HEALTH_OK 2026-03-10T10:30:00.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:00 vm01 ceph-mon[48597]: pgmap v277: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T10:30:03.538 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:03 vm01 ceph-mon[48597]: pgmap v278: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:03.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:03 vm08 ceph-mon[50506]: pgmap v278: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:05.546 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:05 vm08 ceph-mon[50506]: pgmap v279: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:05.548 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:05 vm01 ceph-mon[48597]: pgmap v279: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:05.602 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:30:05.602 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:30:05.642 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:30:05.643 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:30:06.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:06 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:30:06.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:06 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:30:06.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:06 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:30:06.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:06 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:30:06.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:06 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:30:06.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:06 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:30:06.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:06 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:30:06.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:06 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:30:06.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:06 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:30:06.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:06 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:30:07.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:07 vm08 ceph-mon[50506]: pgmap v280: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:30:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:07 vm01 ceph-mon[48597]: pgmap v280: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:30:09.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:09 vm08 ceph-mon[50506]: pgmap v281: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:09.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:09 vm01 ceph-mon[48597]: pgmap v281: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:10.645 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:30:10.646 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:30:10.671 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:30:10.672 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:30:10.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:30:10.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:30:10.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:10 vm08 ceph-mon[50506]: pgmap v282: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:30:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:30:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:10 vm01 ceph-mon[48597]: pgmap v282: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:13.539 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:13 vm01 ceph-mon[48597]: pgmap v283: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:30:13.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:13 vm08 ceph-mon[50506]: pgmap v283: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:30:15.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:15 vm08 ceph-mon[50506]: pgmap v284: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:15.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:15 vm01 ceph-mon[48597]: pgmap v284: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:15.673 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:30:15.674 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:30:15.708 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:30:15.709 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:30:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:17 vm08 ceph-mon[50506]: pgmap v285: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:30:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:17 vm01 ceph-mon[48597]: pgmap v285: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:30:19.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:19 vm08 ceph-mon[50506]: pgmap v286: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:19 vm01 ceph-mon[48597]: pgmap v286: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:20.710 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:30:20.711 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:30:20.738 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:30:20.738 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:30:20.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:20 vm08 ceph-mon[50506]: pgmap v287: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:20.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:20 vm01 ceph-mon[48597]: pgmap v287: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:23.540 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:23 vm01 ceph-mon[48597]: pgmap v288: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:30:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:23 vm08 ceph-mon[50506]: pgmap v288: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:30:25.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:25 vm08 ceph-mon[50506]: pgmap v289: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:25.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:25 vm01 ceph-mon[48597]: pgmap v289: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:25.740 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:30:25.740 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:30:25.766 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:30:25.766 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:30:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:27 vm08 ceph-mon[50506]: pgmap v290: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:30:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:27 vm01 ceph-mon[48597]: pgmap v290: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:30:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:29 vm08 ceph-mon[50506]: pgmap v291: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:29.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:29 vm01 ceph-mon[48597]: pgmap v291: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:30.767 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:30:30.768 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:30:30.793 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:30:30.793 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:30:30.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:30 vm08 ceph-mon[50506]: pgmap v292: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:30 vm01 ceph-mon[48597]: pgmap v292: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:33.540 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:33 vm01 ceph-mon[48597]: pgmap v293: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:30:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:33 vm08 ceph-mon[50506]: pgmap v293: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:30:35.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:35 vm08 ceph-mon[50506]: pgmap v294: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:35.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:35 vm01 ceph-mon[48597]: pgmap v294: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:35.795 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:30:35.795 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:30:35.820 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:30:35.821 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:30:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:37 vm08 ceph-mon[50506]: pgmap v295: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:30:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:37 vm01 ceph-mon[48597]: pgmap v295: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:30:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:39 vm08 ceph-mon[50506]: pgmap v296: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:39.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:39 vm01 ceph-mon[48597]: pgmap v296: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:40.822 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:30:40.822 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:30:40.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:40 vm08 ceph-mon[50506]: pgmap v297: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:40.847 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:30:40.848 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:30:40.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:40 vm01 ceph-mon[48597]: pgmap v297: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:43.541 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:43 vm01 ceph-mon[48597]: pgmap v298: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:30:43.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:43 vm08 ceph-mon[50506]: pgmap v298: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:30:45.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:45 vm08 ceph-mon[50506]: pgmap v299: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:45.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:45 vm01 ceph-mon[48597]: pgmap v299: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:45.849 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:30:45.849 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:30:45.874 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:30:45.875 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:30:47.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:47 vm08 ceph-mon[50506]: pgmap v300: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:30:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:47 vm01 ceph-mon[48597]: pgmap v300: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:30:49.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:49 vm08 ceph-mon[50506]: pgmap v301: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:49.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:49 vm01 ceph-mon[48597]: pgmap v301: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:50.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:50 vm08 ceph-mon[50506]: pgmap v302: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:50.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:50 vm01 ceph-mon[48597]: pgmap v302: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:50.876 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:30:50.877 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:30:50.905 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:30:50.906 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:30:53.541 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:53 vm01 ceph-mon[48597]: pgmap v303: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:30:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:53 vm08 ceph-mon[50506]: pgmap v303: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:30:55.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:55 vm08 ceph-mon[50506]: pgmap v304: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:55 vm01 ceph-mon[48597]: pgmap v304: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:55.907 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:30:55.907 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:30:55.932 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:30:55.932 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:30:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:57 vm08 ceph-mon[50506]: pgmap v305: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:30:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:57 vm01 ceph-mon[48597]: pgmap v305: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:30:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:30:59 vm08 ceph-mon[50506]: pgmap v306: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:30:59.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:30:59 vm01 ceph-mon[48597]: pgmap v306: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:00.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:00 vm08 ceph-mon[50506]: pgmap v307: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:00.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:00 vm01 ceph-mon[48597]: pgmap v307: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:00.934 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:31:00.934 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:31:00.959 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:31:00.960 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:31:03.542 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:03 vm01 ceph-mon[48597]: pgmap v308: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:31:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:03 vm08 ceph-mon[50506]: pgmap v308: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:31:05.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:05 vm08 ceph-mon[50506]: pgmap v309: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:05.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:05 vm01 ceph-mon[48597]: pgmap v309: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:05.961 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:31:05.963 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:31:06.002 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:31:06.002 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:31:06.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:06 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:31:06.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:06 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:31:06.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:06 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:31:06.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:06 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:31:06.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:06 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:31:06.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:06 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:31:06.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:06 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:31:06.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:06 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:31:06.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:06 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:31:06.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:06 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:31:07.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:07 vm08 ceph-mon[50506]: pgmap v310: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:31:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:07 vm01 ceph-mon[48597]: pgmap v310: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:31:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:09 vm08 ceph-mon[50506]: pgmap v311: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:09.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:09 vm01 ceph-mon[48597]: pgmap v311: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:10.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:31:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:31:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:10 vm08 ceph-mon[50506]: pgmap v312: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:31:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:31:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:10 vm01 ceph-mon[48597]: pgmap v312: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:11.003 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:31:11.004 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:31:11.028 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:31:11.029 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:31:13.542 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:13 vm01 ceph-mon[48597]: pgmap v313: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:31:13.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:13 vm08 ceph-mon[50506]: pgmap v313: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:31:15.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:15 vm08 ceph-mon[50506]: pgmap v314: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:15.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:15 vm01 ceph-mon[48597]: pgmap v314: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:16.030 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:31:16.030 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:31:16.055 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:31:16.055 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:31:17.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:17 vm08 ceph-mon[50506]: pgmap v315: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:31:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:17 vm01 ceph-mon[48597]: pgmap v315: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:31:19.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:19 vm08 ceph-mon[50506]: pgmap v316: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:19.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:19 vm01 ceph-mon[48597]: pgmap v316: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:20.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:20 vm08 ceph-mon[50506]: pgmap v317: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:20.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:20 vm01 ceph-mon[48597]: pgmap v317: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:21.057 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:31:21.057 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:31:21.083 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:31:21.084 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:31:23.543 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:23 vm01 ceph-mon[48597]: pgmap v318: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:31:23.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:23 vm08 ceph-mon[50506]: pgmap v318: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:31:25.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:25 vm08 ceph-mon[50506]: pgmap v319: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:25.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:25 vm01 ceph-mon[48597]: pgmap v319: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:26.085 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:31:26.086 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:31:26.112 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:31:26.112 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:31:27.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:27 vm08 ceph-mon[50506]: pgmap v320: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:31:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:27 vm01 ceph-mon[48597]: pgmap v320: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:31:29.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:29 vm08 ceph-mon[50506]: pgmap v321: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:29.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:29 vm01 ceph-mon[48597]: pgmap v321: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:30.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:30 vm08 ceph-mon[50506]: pgmap v322: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:30 vm01 ceph-mon[48597]: pgmap v322: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:31.113 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:31:31.114 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:31:31.143 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:31:31.144 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:31:33.543 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:33 vm01 ceph-mon[48597]: pgmap v323: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:31:33.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:33 vm08 ceph-mon[50506]: pgmap v323: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:31:35.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:35 vm08 ceph-mon[50506]: pgmap v324: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:35.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:35 vm01 ceph-mon[48597]: pgmap v324: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:36.145 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:31:36.145 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:31:36.171 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:31:36.171 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:31:37.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:37 vm08 ceph-mon[50506]: pgmap v325: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:31:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:37 vm01 ceph-mon[48597]: pgmap v325: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:31:39.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:39 vm08 ceph-mon[50506]: pgmap v326: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:39.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:39 vm01 ceph-mon[48597]: pgmap v326: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:40.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:40 vm08 ceph-mon[50506]: pgmap v327: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:40.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:40 vm01 ceph-mon[48597]: pgmap v327: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:41.173 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:31:41.173 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:31:41.199 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:31:41.199 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:31:43.544 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:43 vm01 ceph-mon[48597]: pgmap v328: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:31:43.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:43 vm08 ceph-mon[50506]: pgmap v328: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:31:45.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:45 vm08 ceph-mon[50506]: pgmap v329: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:45.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:45 vm01 ceph-mon[48597]: pgmap v329: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:46.200 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:31:46.201 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:31:46.226 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:31:46.227 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:31:47.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:47 vm08 ceph-mon[50506]: pgmap v330: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:31:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:47 vm01 ceph-mon[48597]: pgmap v330: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:31:49.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:49 vm08 ceph-mon[50506]: pgmap v331: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:49.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:49 vm01 ceph-mon[48597]: pgmap v331: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:50.787 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:50 vm01 ceph-mon[48597]: pgmap v332: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:50.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:50 vm08 ceph-mon[50506]: pgmap v332: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:51.228 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:31:51.228 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:31:51.253 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:31:51.254 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:31:53.546 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:53 vm01 ceph-mon[48597]: pgmap v333: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:31:53.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:53 vm08 ceph-mon[50506]: pgmap v333: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:31:55.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:55 vm08 ceph-mon[50506]: pgmap v334: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:55 vm01 ceph-mon[48597]: pgmap v334: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:56.255 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:31:56.255 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:31:56.281 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:31:56.281 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:31:57.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:57 vm08 ceph-mon[50506]: pgmap v335: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:31:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:57 vm01 ceph-mon[48597]: pgmap v335: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:31:59.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:31:59 vm08 ceph-mon[50506]: pgmap v336: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:31:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:31:59 vm01 ceph-mon[48597]: pgmap v336: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:00.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:00 vm08 ceph-mon[50506]: pgmap v337: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:00.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:00 vm01 ceph-mon[48597]: pgmap v337: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:01.283 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:32:01.283 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:32:01.309 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:32:01.309 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:32:03.547 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:03 vm01 ceph-mon[48597]: pgmap v338: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:32:03.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:03 vm08 ceph-mon[50506]: pgmap v338: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:32:05.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:05 vm08 ceph-mon[50506]: pgmap v339: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:05.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:05 vm01 ceph-mon[48597]: pgmap v339: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:06.278 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:06 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:32:06.278 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:06 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:32:06.278 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:06 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:32:06.311 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:32:06.311 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:32:06.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:06 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:32:06.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:06 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:32:06.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:06 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:32:06.345 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:32:06.346 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:32:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:07 vm08 ceph-mon[50506]: pgmap v340: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:32:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:07 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:32:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:07 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:32:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:07 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:32:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:07 vm01 ceph-mon[48597]: pgmap v340: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:32:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:07 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:32:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:07 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:32:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:07 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:32:09.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:09 vm08 ceph-mon[50506]: pgmap v341: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:09.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:09 vm01 ceph-mon[48597]: pgmap v341: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:10.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:32:10.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:32:10.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:10 vm08 ceph-mon[50506]: pgmap v342: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:32:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:32:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:10 vm01 ceph-mon[48597]: pgmap v342: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:11.350 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:32:11.351 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:32:11.376 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:32:11.376 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:32:13.548 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:13 vm01 ceph-mon[48597]: pgmap v343: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:32:13.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:13 vm08 ceph-mon[50506]: pgmap v343: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:32:15.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:15 vm08 ceph-mon[50506]: pgmap v344: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:15.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:15 vm01 ceph-mon[48597]: pgmap v344: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:16.378 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:32:16.378 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:32:16.406 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:32:16.407 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:32:17.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:17 vm08 ceph-mon[50506]: pgmap v345: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:32:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:17 vm01 ceph-mon[48597]: pgmap v345: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:32:19.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:19 vm08 ceph-mon[50506]: pgmap v346: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:19 vm01 ceph-mon[48597]: pgmap v346: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:20.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:20 vm08 ceph-mon[50506]: pgmap v347: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:20.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:20 vm01 ceph-mon[48597]: pgmap v347: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:21.408 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:32:21.409 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:32:21.435 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:32:21.436 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:32:23.549 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:23 vm01 ceph-mon[48597]: pgmap v348: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:32:23.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:23 vm08 ceph-mon[50506]: pgmap v348: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:32:25.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:25 vm08 ceph-mon[50506]: pgmap v349: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:25.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:25 vm01 ceph-mon[48597]: pgmap v349: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:26.437 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:32:26.438 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:32:26.463 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:32:26.464 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:32:27.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:27 vm08 ceph-mon[50506]: pgmap v350: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:32:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:27 vm01 ceph-mon[48597]: pgmap v350: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:32:29.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:29 vm08 ceph-mon[50506]: pgmap v351: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:29.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:29 vm01 ceph-mon[48597]: pgmap v351: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:30.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:30 vm08 ceph-mon[50506]: pgmap v352: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:30 vm01 ceph-mon[48597]: pgmap v352: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:31.465 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:32:31.465 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:32:31.490 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:32:31.490 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:32:33.550 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:33 vm01 ceph-mon[48597]: pgmap v353: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:32:33.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:33 vm08 ceph-mon[50506]: pgmap v353: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:32:35.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:35 vm08 ceph-mon[50506]: pgmap v354: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:35.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:35 vm01 ceph-mon[48597]: pgmap v354: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:36.491 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:32:36.492 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:32:36.517 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:32:36.518 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:32:37.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:37 vm08 ceph-mon[50506]: pgmap v355: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:32:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:37 vm01 ceph-mon[48597]: pgmap v355: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:32:39.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:39 vm08 ceph-mon[50506]: pgmap v356: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:39.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:39 vm01 ceph-mon[48597]: pgmap v356: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:40.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:40 vm01 ceph-mon[48597]: pgmap v357: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:41.093 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:40 vm08 ceph-mon[50506]: pgmap v357: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:41.519 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:32:41.519 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:32:41.545 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:32:41.545 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:32:43.551 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:43 vm01 ceph-mon[48597]: pgmap v358: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:32:43.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:43 vm08 ceph-mon[50506]: pgmap v358: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:32:45.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:45 vm08 ceph-mon[50506]: pgmap v359: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:45.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:45 vm01 ceph-mon[48597]: pgmap v359: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:46.547 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:32:46.547 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:32:46.573 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:32:46.574 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:32:47.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:47 vm08 ceph-mon[50506]: pgmap v360: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:32:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:47 vm01 ceph-mon[48597]: pgmap v360: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:32:49.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:49 vm08 ceph-mon[50506]: pgmap v361: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:49.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:49 vm01 ceph-mon[48597]: pgmap v361: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:50.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:50 vm01 ceph-mon[48597]: pgmap v362: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:51.093 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:50 vm08 ceph-mon[50506]: pgmap v362: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:51.575 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:32:51.575 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:32:51.600 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:32:51.601 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:32:53.553 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:53 vm01 ceph-mon[48597]: pgmap v363: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:32:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:53 vm08 ceph-mon[50506]: pgmap v363: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:32:55.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:55 vm08 ceph-mon[50506]: pgmap v364: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:55 vm01 ceph-mon[48597]: pgmap v364: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:56.602 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:32:56.603 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:32:56.628 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:32:56.629 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:32:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:57 vm08 ceph-mon[50506]: pgmap v365: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:32:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:57 vm01 ceph-mon[48597]: pgmap v365: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:32:59.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:32:59 vm08 ceph-mon[50506]: pgmap v366: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:32:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:32:59 vm01 ceph-mon[48597]: pgmap v366: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:00.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:00 vm01 ceph-mon[48597]: pgmap v367: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:01.093 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:00 vm08 ceph-mon[50506]: pgmap v367: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:01.630 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:33:01.631 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:33:01.655 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:33:01.656 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:33:03.553 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:03 vm01 ceph-mon[48597]: pgmap v368: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:33:03.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:03 vm08 ceph-mon[50506]: pgmap v368: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:33:05.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:05 vm08 ceph-mon[50506]: pgmap v369: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:05.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:05 vm01 ceph-mon[48597]: pgmap v369: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:06.657 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:33:06.658 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:33:06.690 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:33:06.690 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:33:07.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:07 vm08 ceph-mon[50506]: pgmap v370: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:33:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:07 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:33:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:07 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:33:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:07 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:33:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:07 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:33:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:07 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:33:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:07 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:33:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:07 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:33:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:07 vm01 ceph-mon[48597]: pgmap v370: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:33:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:07 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:33:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:07 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:33:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:07 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:33:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:07 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:33:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:07 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:33:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:07 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:33:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:07 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:33:09.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:09 vm08 ceph-mon[50506]: pgmap v371: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:09.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:09 vm01 ceph-mon[48597]: pgmap v371: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:33:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:33:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:10 vm01 ceph-mon[48597]: pgmap v372: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:33:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:33:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:10 vm08 ceph-mon[50506]: pgmap v372: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:11.691 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:33:11.692 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:33:11.718 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:33:11.718 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:33:13.555 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:13 vm01 ceph-mon[48597]: pgmap v373: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:33:13.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:13 vm08 ceph-mon[50506]: pgmap v373: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:33:15.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:15 vm08 ceph-mon[50506]: pgmap v374: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:15.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:15 vm01 ceph-mon[48597]: pgmap v374: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:16.720 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:33:16.720 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:33:16.746 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:33:16.746 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:33:17.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:17 vm08 ceph-mon[50506]: pgmap v375: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:33:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:17 vm01 ceph-mon[48597]: pgmap v375: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:33:19.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:19 vm08 ceph-mon[50506]: pgmap v376: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:19 vm01 ceph-mon[48597]: pgmap v376: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:20.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:20 vm01 ceph-mon[48597]: pgmap v377: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:21.093 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:20 vm08 ceph-mon[50506]: pgmap v377: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:21.748 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:33:21.748 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:33:21.774 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:33:21.775 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:33:23.556 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:23 vm01 ceph-mon[48597]: pgmap v378: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:33:23.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:23 vm08 ceph-mon[50506]: pgmap v378: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:33:25.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:25 vm08 ceph-mon[50506]: pgmap v379: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:25.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:25 vm01 ceph-mon[48597]: pgmap v379: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:26.776 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:33:26.776 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:33:26.803 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:33:26.803 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:33:27.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:27 vm08 ceph-mon[50506]: pgmap v380: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:33:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:27 vm01 ceph-mon[48597]: pgmap v380: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:33:29.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:29 vm08 ceph-mon[50506]: pgmap v381: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:29.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:29 vm01 ceph-mon[48597]: pgmap v381: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:30 vm01 ceph-mon[48597]: pgmap v382: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:31.093 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:30 vm08 ceph-mon[50506]: pgmap v382: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:31.805 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:33:31.805 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:33:31.832 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:33:31.832 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:33:33.556 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:33 vm01 ceph-mon[48597]: pgmap v383: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:33:33.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:33 vm08 ceph-mon[50506]: pgmap v383: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:33:35.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:35 vm08 ceph-mon[50506]: pgmap v384: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:35.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:35 vm01 ceph-mon[48597]: pgmap v384: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:36.834 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:33:36.834 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:33:36.860 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:33:36.860 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:33:37.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:37 vm08 ceph-mon[50506]: pgmap v385: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:33:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:37 vm01 ceph-mon[48597]: pgmap v385: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:33:39.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:39 vm01 ceph-mon[48597]: pgmap v386: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:39.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:39 vm08 ceph-mon[50506]: pgmap v386: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:40.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:40 vm01 ceph-mon[48597]: pgmap v387: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:41.093 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:40 vm08 ceph-mon[50506]: pgmap v387: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:41.862 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:33:41.862 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:33:41.888 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:33:41.889 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:33:43.558 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:43 vm01 ceph-mon[48597]: pgmap v388: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:33:43.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:43 vm08 ceph-mon[50506]: pgmap v388: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:33:45.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:45 vm08 ceph-mon[50506]: pgmap v389: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:45.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:45 vm01 ceph-mon[48597]: pgmap v389: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:46.890 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:33:46.891 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:33:46.915 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:33:46.916 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:33:47.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:47 vm08 ceph-mon[50506]: pgmap v390: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:33:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:47 vm01 ceph-mon[48597]: pgmap v390: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:33:49.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:49 vm08 ceph-mon[50506]: pgmap v391: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:49.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:49 vm01 ceph-mon[48597]: pgmap v391: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:50.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:50 vm01 ceph-mon[48597]: pgmap v392: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:51.093 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:50 vm08 ceph-mon[50506]: pgmap v392: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:51.917 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:33:51.917 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:33:51.942 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:33:51.942 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:33:53.559 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:53 vm01 ceph-mon[48597]: pgmap v393: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:33:53.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:53 vm08 ceph-mon[50506]: pgmap v393: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:33:55.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:55 vm08 ceph-mon[50506]: pgmap v394: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:55 vm01 ceph-mon[48597]: pgmap v394: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:56.944 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:33:56.944 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:33:56.972 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:33:56.973 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:33:57.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:57 vm08 ceph-mon[50506]: pgmap v395: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:33:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:57 vm01 ceph-mon[48597]: pgmap v395: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:33:59.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:33:59 vm08 ceph-mon[50506]: pgmap v396: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:33:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:33:59 vm01 ceph-mon[48597]: pgmap v396: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:00.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:00 vm01 ceph-mon[48597]: pgmap v397: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:01.093 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:00 vm08 ceph-mon[50506]: pgmap v397: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:01.974 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:34:01.974 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:34:01.999 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:34:02.000 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:34:03.559 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:03 vm01 ceph-mon[48597]: pgmap v398: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:34:03.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:03 vm08 ceph-mon[50506]: pgmap v398: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:34:05.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:05 vm08 ceph-mon[50506]: pgmap v399: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:05.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:05 vm01 ceph-mon[48597]: pgmap v399: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:07.001 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:34:07.002 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:34:07.042 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:34:07.042 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:34:07.533 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:07 vm01 ceph-mon[48597]: pgmap v400: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:34:07.533 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:07 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:34:07.533 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:07 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:34:07.533 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:07 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:34:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:07 vm08 ceph-mon[50506]: pgmap v400: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:34:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:07 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:34:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:07 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:34:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:07 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:34:09.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:08 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:34:09.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:08 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:34:09.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:08 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:34:09.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:08 vm08 ceph-mon[50506]: pgmap v401: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:09.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:08 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:34:09.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:08 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:34:09.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:08 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:34:09.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:08 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:34:09.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:08 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:34:09.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:08 vm01 ceph-mon[48597]: pgmap v401: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:09.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:08 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:34:09.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:08 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:34:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:34:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:34:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:10 vm01 ceph-mon[48597]: pgmap v402: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:34:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:34:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:10 vm08 ceph-mon[50506]: pgmap v402: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:12.043 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:34:12.044 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:34:12.071 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:34:12.071 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:34:13.560 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:13 vm01 ceph-mon[48597]: pgmap v403: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:34:13.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:13 vm08 ceph-mon[50506]: pgmap v403: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:34:15.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:15 vm08 ceph-mon[50506]: pgmap v404: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:15.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:15 vm01 ceph-mon[48597]: pgmap v404: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:17.073 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:34:17.074 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:34:17.100 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:34:17.101 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:34:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:17 vm08 ceph-mon[50506]: pgmap v405: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:34:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:17 vm01 ceph-mon[48597]: pgmap v405: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:34:19.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:19 vm08 ceph-mon[50506]: pgmap v406: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:19 vm01 ceph-mon[48597]: pgmap v406: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:20.992 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:20 vm01 ceph-mon[48597]: pgmap v407: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:21.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:20 vm08 ceph-mon[50506]: pgmap v407: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:22.102 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:34:22.103 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:34:22.129 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:34:22.129 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:34:23.561 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:23 vm01 ceph-mon[48597]: pgmap v408: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:34:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:23 vm08 ceph-mon[50506]: pgmap v408: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:34:25.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:25 vm08 ceph-mon[50506]: pgmap v409: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:25.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:25 vm01 ceph-mon[48597]: pgmap v409: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:27.131 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:34:27.131 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:34:27.157 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:34:27.157 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:34:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:27 vm08 ceph-mon[50506]: pgmap v410: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:34:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:27 vm01 ceph-mon[48597]: pgmap v410: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:34:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:29 vm08 ceph-mon[50506]: pgmap v411: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:29.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:29 vm01 ceph-mon[48597]: pgmap v411: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:30.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:30 vm01 ceph-mon[48597]: pgmap v412: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:31.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:30 vm08 ceph-mon[50506]: pgmap v412: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:32.159 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:34:32.159 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:34:32.184 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:34:32.184 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:34:33.562 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:33 vm01 ceph-mon[48597]: pgmap v413: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:34:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:33 vm08 ceph-mon[50506]: pgmap v413: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:34:35.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:35 vm08 ceph-mon[50506]: pgmap v414: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:35.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:35 vm01 ceph-mon[48597]: pgmap v414: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:37.185 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:34:37.186 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:34:37.213 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:34:37.214 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:34:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:37 vm08 ceph-mon[50506]: pgmap v415: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:34:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:37 vm01 ceph-mon[48597]: pgmap v415: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:34:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:39 vm08 ceph-mon[50506]: pgmap v416: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:39.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:39 vm01 ceph-mon[48597]: pgmap v416: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:40.992 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:40 vm01 ceph-mon[48597]: pgmap v417: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:41.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:40 vm08 ceph-mon[50506]: pgmap v417: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:42.215 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:34:42.215 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:34:42.242 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:34:42.242 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:34:43.563 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:43 vm01 ceph-mon[48597]: pgmap v418: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:34:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:43 vm08 ceph-mon[50506]: pgmap v418: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:34:45.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:45 vm08 ceph-mon[50506]: pgmap v419: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:45.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:45 vm01 ceph-mon[48597]: pgmap v419: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:47.244 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:34:47.244 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:34:47.272 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:34:47.272 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:34:47.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:47 vm08 ceph-mon[50506]: pgmap v420: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:34:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:47 vm01 ceph-mon[48597]: pgmap v420: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:34:49.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:49 vm08 ceph-mon[50506]: pgmap v421: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:49.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:49 vm01 ceph-mon[48597]: pgmap v421: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:51.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:50 vm08 ceph-mon[50506]: pgmap v422: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:51.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:50 vm01 ceph-mon[48597]: pgmap v422: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:52.274 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:34:52.274 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:34:52.299 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:34:52.300 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:34:53.563 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:53 vm01 ceph-mon[48597]: pgmap v423: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:34:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:53 vm08 ceph-mon[50506]: pgmap v423: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:34:55.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:55 vm08 ceph-mon[50506]: pgmap v424: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:55.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:55 vm01 ceph-mon[48597]: pgmap v424: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:57.301 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:34:57.302 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:34:57.328 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:34:57.329 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:34:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:57 vm08 ceph-mon[50506]: pgmap v425: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:34:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:57 vm01 ceph-mon[48597]: pgmap v425: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:34:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:34:59 vm08 ceph-mon[50506]: pgmap v426: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:34:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:34:59 vm01 ceph-mon[48597]: pgmap v426: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:00.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:00 vm01 ceph-mon[48597]: pgmap v427: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:01.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:00 vm08 ceph-mon[50506]: pgmap v427: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:02.330 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:35:02.331 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:35:02.357 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:35:02.358 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:35:03.564 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:03 vm01 ceph-mon[48597]: pgmap v428: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:35:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:03 vm08 ceph-mon[50506]: pgmap v428: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:35:05.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:05 vm08 ceph-mon[50506]: pgmap v429: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:05.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:05 vm01 ceph-mon[48597]: pgmap v429: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:07.359 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:35:07.360 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:35:07.385 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:35:07.386 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:35:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:07 vm08 ceph-mon[50506]: pgmap v430: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:35:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:07 vm01 ceph-mon[48597]: pgmap v430: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:35:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:09 vm08 ceph-mon[50506]: pgmap v431: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:09 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:35:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:09 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:35:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:09 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:35:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:09 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:35:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:09 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:35:09.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:09 vm01 ceph-mon[48597]: pgmap v431: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:09.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:09 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:35:09.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:09 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:35:09.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:09 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:35:09.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:09 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:35:09.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:09 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:35:10.992 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:35:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:35:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:10 vm01 ceph-mon[48597]: pgmap v432: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:35:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:35:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:10 vm08 ceph-mon[50506]: pgmap v432: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:12.387 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:35:12.388 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:35:12.413 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:35:12.414 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:35:13.565 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:13 vm01 ceph-mon[48597]: pgmap v433: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:35:13.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:13 vm08 ceph-mon[50506]: pgmap v433: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:35:15.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:15 vm08 ceph-mon[50506]: pgmap v434: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:15.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:15 vm01 ceph-mon[48597]: pgmap v434: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:17.415 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:35:17.416 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:35:17.440 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:35:17.441 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:35:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:17 vm08 ceph-mon[50506]: pgmap v435: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:35:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:17 vm01 ceph-mon[48597]: pgmap v435: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:35:19.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:19 vm08 ceph-mon[50506]: pgmap v436: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:19 vm01 ceph-mon[48597]: pgmap v436: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:20.992 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:20 vm01 ceph-mon[48597]: pgmap v437: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:21.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:20 vm08 ceph-mon[50506]: pgmap v437: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:22.442 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:35:22.443 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:35:22.468 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:35:22.468 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:35:23.565 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:23 vm01 ceph-mon[48597]: pgmap v438: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:35:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:23 vm08 ceph-mon[50506]: pgmap v438: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:35:25.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:25 vm08 ceph-mon[50506]: pgmap v439: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:25.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:25 vm01 ceph-mon[48597]: pgmap v439: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:27.470 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:35:27.470 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:35:27.496 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:35:27.496 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:35:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:27 vm08 ceph-mon[50506]: pgmap v440: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:35:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:27 vm01 ceph-mon[48597]: pgmap v440: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:35:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:29 vm08 ceph-mon[50506]: pgmap v441: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:29.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:29 vm01 ceph-mon[48597]: pgmap v441: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:30.992 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:30 vm01 ceph-mon[48597]: pgmap v442: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:31.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:30 vm08 ceph-mon[50506]: pgmap v442: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:32.498 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:35:32.498 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:35:32.524 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:35:32.525 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:35:33.566 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:33 vm01 ceph-mon[48597]: pgmap v443: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:35:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:33 vm08 ceph-mon[50506]: pgmap v443: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:35:35.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:35 vm08 ceph-mon[50506]: pgmap v444: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:35.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:35 vm01 ceph-mon[48597]: pgmap v444: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:37.526 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:35:37.526 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:35:37.551 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:35:37.552 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:35:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:37 vm08 ceph-mon[50506]: pgmap v445: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:35:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:37 vm01 ceph-mon[48597]: pgmap v445: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:35:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:39 vm08 ceph-mon[50506]: pgmap v446: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:39.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:39 vm01 ceph-mon[48597]: pgmap v446: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:40.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:40 vm01 ceph-mon[48597]: pgmap v447: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:41.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:40 vm08 ceph-mon[50506]: pgmap v447: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:42.553 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:35:42.553 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:35:42.579 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:35:42.579 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:35:43.567 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:43 vm01 ceph-mon[48597]: pgmap v448: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:35:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:43 vm08 ceph-mon[50506]: pgmap v448: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:35:45.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:45 vm08 ceph-mon[50506]: pgmap v449: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:45.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:45 vm01 ceph-mon[48597]: pgmap v449: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:47.581 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:35:47.581 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:35:47.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:47 vm08 ceph-mon[50506]: pgmap v450: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:35:47.606 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:35:47.607 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:35:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:47 vm01 ceph-mon[48597]: pgmap v450: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:35:49.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:49 vm08 ceph-mon[50506]: pgmap v451: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:49.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:49 vm01 ceph-mon[48597]: pgmap v451: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:51.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:50 vm08 ceph-mon[50506]: pgmap v452: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:51.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:50 vm01 ceph-mon[48597]: pgmap v452: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:52.608 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:35:52.609 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:35:52.634 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:35:52.635 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:35:53.568 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:53 vm01 ceph-mon[48597]: pgmap v453: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:35:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:53 vm08 ceph-mon[50506]: pgmap v453: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:35:55.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:55 vm08 ceph-mon[50506]: pgmap v454: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:55 vm01 ceph-mon[48597]: pgmap v454: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:57 vm08 ceph-mon[50506]: pgmap v455: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:35:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:57 vm01 ceph-mon[48597]: pgmap v455: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:35:57.636 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:35:57.637 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:35:57.662 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:35:57.663 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:35:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:35:59 vm08 ceph-mon[50506]: pgmap v456: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:35:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:35:59 vm01 ceph-mon[48597]: pgmap v456: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:00.992 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:00 vm01 ceph-mon[48597]: pgmap v457: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:01.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:00 vm08 ceph-mon[50506]: pgmap v457: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:02.664 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:36:02.665 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:36:02.696 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:36:02.697 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:36:03.569 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:03 vm01 ceph-mon[48597]: pgmap v458: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:36:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:03 vm08 ceph-mon[50506]: pgmap v458: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:36:05.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:05 vm08 ceph-mon[50506]: pgmap v459: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:05.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:05 vm01 ceph-mon[48597]: pgmap v459: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:07 vm08 ceph-mon[50506]: pgmap v460: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:36:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:07 vm01 ceph-mon[48597]: pgmap v460: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:36:07.698 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:36:07.699 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:36:07.724 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:36:07.725 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:36:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:09 vm08 ceph-mon[50506]: pgmap v461: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:09 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:36:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:09 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:36:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:09 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:36:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:09 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:36:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:09 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:36:09.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:09 vm01 ceph-mon[48597]: pgmap v461: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:09.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:09 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:36:09.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:09 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:36:09.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:09 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:36:09.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:09 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:36:09.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:09 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:36:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:36:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:36:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:10 vm01 ceph-mon[48597]: pgmap v462: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:36:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:36:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:10 vm08 ceph-mon[50506]: pgmap v462: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:12.726 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:36:12.727 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:36:12.754 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:36:12.754 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:36:13.570 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:13 vm01 ceph-mon[48597]: pgmap v463: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:36:13.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:13 vm08 ceph-mon[50506]: pgmap v463: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:36:15.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:15 vm08 ceph-mon[50506]: pgmap v464: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:15.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:15 vm01 ceph-mon[48597]: pgmap v464: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:17 vm08 ceph-mon[50506]: pgmap v465: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:36:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:17 vm01 ceph-mon[48597]: pgmap v465: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:36:17.756 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:36:17.756 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:36:17.783 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:36:17.784 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:36:19.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:19 vm08 ceph-mon[50506]: pgmap v466: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:19 vm01 ceph-mon[48597]: pgmap v466: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:20.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:20 vm01 ceph-mon[48597]: pgmap v467: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:21.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:20 vm08 ceph-mon[50506]: pgmap v467: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:22.785 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:36:22.785 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:36:22.813 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:36:22.814 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:36:23.571 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:23 vm01 ceph-mon[48597]: pgmap v468: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:36:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:23 vm08 ceph-mon[50506]: pgmap v468: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:36:25.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:25 vm08 ceph-mon[50506]: pgmap v469: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:25.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:25 vm01 ceph-mon[48597]: pgmap v469: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:27 vm08 ceph-mon[50506]: pgmap v470: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:36:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:27 vm01 ceph-mon[48597]: pgmap v470: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:36:27.815 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:36:27.816 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:36:27.843 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:36:27.844 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:36:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:29 vm08 ceph-mon[50506]: pgmap v471: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:29.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:29 vm01 ceph-mon[48597]: pgmap v471: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:30.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:30 vm01 ceph-mon[48597]: pgmap v472: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:31.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:30 vm08 ceph-mon[50506]: pgmap v472: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:32.845 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:36:32.846 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:36:32.872 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:36:32.873 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:36:33.571 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:33 vm01 ceph-mon[48597]: pgmap v473: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:36:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:33 vm08 ceph-mon[50506]: pgmap v473: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:36:35.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:35 vm08 ceph-mon[50506]: pgmap v474: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:35.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:35 vm01 ceph-mon[48597]: pgmap v474: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:37 vm08 ceph-mon[50506]: pgmap v475: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:36:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:37 vm01 ceph-mon[48597]: pgmap v475: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:36:37.874 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:36:37.875 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:36:37.900 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:36:37.900 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:36:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:39 vm08 ceph-mon[50506]: pgmap v476: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:39.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:39 vm01 ceph-mon[48597]: pgmap v476: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:40.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:40 vm01 ceph-mon[48597]: pgmap v477: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:41.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:40 vm08 ceph-mon[50506]: pgmap v477: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:42.902 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:36:42.902 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:36:42.927 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:36:42.928 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:36:43.572 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:43 vm01 ceph-mon[48597]: pgmap v478: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:36:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:43 vm08 ceph-mon[50506]: pgmap v478: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:36:45.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:45 vm08 ceph-mon[50506]: pgmap v479: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:45.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:45 vm01 ceph-mon[48597]: pgmap v479: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:47.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:47 vm08 ceph-mon[50506]: pgmap v480: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:36:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:47 vm01 ceph-mon[48597]: pgmap v480: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:36:47.929 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:36:47.930 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:36:47.955 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:36:47.956 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:36:49.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:49 vm08 ceph-mon[50506]: pgmap v481: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:49.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:49 vm01 ceph-mon[48597]: pgmap v481: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:50.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:50 vm01 ceph-mon[48597]: pgmap v482: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:51.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:50 vm08 ceph-mon[50506]: pgmap v482: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:52.957 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:36:52.958 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:36:52.985 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:36:52.985 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:36:53.574 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:53 vm01 ceph-mon[48597]: pgmap v483: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:36:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:53 vm08 ceph-mon[50506]: pgmap v483: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:36:55.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:55 vm08 ceph-mon[50506]: pgmap v484: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:55 vm01 ceph-mon[48597]: pgmap v484: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:57 vm08 ceph-mon[50506]: pgmap v485: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:36:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:57 vm01 ceph-mon[48597]: pgmap v485: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:36:57.987 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:36:57.987 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:36:58.014 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:36:58.015 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:36:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:36:59 vm08 ceph-mon[50506]: pgmap v486: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:36:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:36:59 vm01 ceph-mon[48597]: pgmap v486: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:00.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:00 vm01 ceph-mon[48597]: pgmap v487: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:01.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:00 vm08 ceph-mon[50506]: pgmap v487: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:03.016 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:37:03.017 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:37:03.042 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:37:03.043 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:37:03.575 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:03 vm01 ceph-mon[48597]: pgmap v488: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:37:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:03 vm08 ceph-mon[50506]: pgmap v488: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:37:05.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:05 vm08 ceph-mon[50506]: pgmap v489: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:05.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:05 vm01 ceph-mon[48597]: pgmap v489: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:07 vm08 ceph-mon[50506]: pgmap v490: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:37:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:07 vm01 ceph-mon[48597]: pgmap v490: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:37:08.044 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:37:08.044 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:37:08.069 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:37:08.069 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:37:09.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:09 vm01 ceph-mon[48597]: pgmap v491: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:09.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:09 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:37:09.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:09 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:37:09.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:09 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:37:09.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:09 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:37:09.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:09 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:37:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:09 vm08 ceph-mon[50506]: pgmap v491: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:09 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:37:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:09 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:37:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:09 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:37:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:09 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:37:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:09 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:37:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:37:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:37:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:10 vm01 ceph-mon[48597]: pgmap v492: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:37:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:37:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:10 vm08 ceph-mon[50506]: pgmap v492: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:13.071 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:37:13.071 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:37:13.099 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:37:13.100 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:37:13.576 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:13 vm01 ceph-mon[48597]: pgmap v493: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:37:13.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:13 vm08 ceph-mon[50506]: pgmap v493: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:37:15.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:15 vm08 ceph-mon[50506]: pgmap v494: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:15.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:15 vm01 ceph-mon[48597]: pgmap v494: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:17 vm08 ceph-mon[50506]: pgmap v495: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:37:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:17 vm01 ceph-mon[48597]: pgmap v495: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:37:18.101 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:37:18.102 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:37:18.132 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:37:18.133 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:37:19.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:19 vm08 ceph-mon[50506]: pgmap v496: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:19 vm01 ceph-mon[48597]: pgmap v496: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:20.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:20 vm01 ceph-mon[48597]: pgmap v497: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:21.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:20 vm08 ceph-mon[50506]: pgmap v497: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:23.134 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:37:23.134 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:37:23.181 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:37:23.181 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:37:23.576 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:23 vm01 ceph-mon[48597]: pgmap v498: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:37:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:23 vm08 ceph-mon[50506]: pgmap v498: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:37:25.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:25 vm08 ceph-mon[50506]: pgmap v499: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:25.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:25 vm01 ceph-mon[48597]: pgmap v499: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:27 vm08 ceph-mon[50506]: pgmap v500: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:37:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:27 vm01 ceph-mon[48597]: pgmap v500: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:37:28.183 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:37:28.183 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:37:28.208 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:37:28.209 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:37:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:29 vm08 ceph-mon[50506]: pgmap v501: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:29.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:29 vm01 ceph-mon[48597]: pgmap v501: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:30.992 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:30 vm01 ceph-mon[48597]: pgmap v502: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:31.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:30 vm08 ceph-mon[50506]: pgmap v502: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:33.210 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:37:33.211 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:37:33.235 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:37:33.236 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:37:33.578 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:33 vm01 ceph-mon[48597]: pgmap v503: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:37:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:33 vm08 ceph-mon[50506]: pgmap v503: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:37:35.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:35 vm08 ceph-mon[50506]: pgmap v504: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:35.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:35 vm01 ceph-mon[48597]: pgmap v504: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:37.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:37 vm08 ceph-mon[50506]: pgmap v505: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:37:37.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:37 vm01 ceph-mon[48597]: pgmap v505: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:37:38.237 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:37:38.238 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:37:38.263 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:37:38.264 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:37:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:39 vm08 ceph-mon[50506]: pgmap v506: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:39.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:39 vm01 ceph-mon[48597]: pgmap v506: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:40.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:40 vm01 ceph-mon[48597]: pgmap v507: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:41.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:40 vm08 ceph-mon[50506]: pgmap v507: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:43.265 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:37:43.265 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:37:43.291 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:37:43.291 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:37:43.578 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:43 vm01 ceph-mon[48597]: pgmap v508: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:37:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:43 vm08 ceph-mon[50506]: pgmap v508: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:37:45.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:45 vm08 ceph-mon[50506]: pgmap v509: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:45.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:45 vm01 ceph-mon[48597]: pgmap v509: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:47.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:47 vm08 ceph-mon[50506]: pgmap v510: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:37:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:47 vm01 ceph-mon[48597]: pgmap v510: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:37:48.292 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:37:48.293 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:37:48.323 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:37:48.323 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:37:49.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:49 vm08 ceph-mon[50506]: pgmap v511: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:49.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:49 vm01 ceph-mon[48597]: pgmap v511: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:50.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:50 vm01 ceph-mon[48597]: pgmap v512: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:51.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:50 vm08 ceph-mon[50506]: pgmap v512: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:53.325 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:37:53.325 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:37:53.352 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:37:53.352 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:37:53.579 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:53 vm01 ceph-mon[48597]: pgmap v513: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:37:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:53 vm08 ceph-mon[50506]: pgmap v513: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:37:55.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:55 vm08 ceph-mon[50506]: pgmap v514: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:55 vm01 ceph-mon[48597]: pgmap v514: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:57 vm08 ceph-mon[50506]: pgmap v515: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:37:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:57 vm01 ceph-mon[48597]: pgmap v515: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:37:58.354 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:37:58.354 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:37:58.380 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:37:58.381 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:37:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:37:59 vm08 ceph-mon[50506]: pgmap v516: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:37:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:37:59 vm01 ceph-mon[48597]: pgmap v516: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:00.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:00 vm01 ceph-mon[48597]: pgmap v517: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:01.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:00 vm08 ceph-mon[50506]: pgmap v517: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:03.382 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:38:03.383 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:38:03.409 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:38:03.409 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:38:03.581 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:03 vm01 ceph-mon[48597]: pgmap v518: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:38:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:03 vm08 ceph-mon[50506]: pgmap v518: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:38:05.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:05 vm08 ceph-mon[50506]: pgmap v519: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:05.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:05 vm01 ceph-mon[48597]: pgmap v519: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:07 vm01 ceph-mon[48597]: pgmap v520: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:38:07.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:07 vm08 ceph-mon[50506]: pgmap v520: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:38:08.411 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:38:08.411 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:38:08.437 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:38:08.437 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:38:09.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:09 vm08 ceph-mon[50506]: pgmap v521: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:09.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:09 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:38:09.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:09 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:38:09.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:09 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:38:09.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:09 vm01 ceph-mon[48597]: pgmap v521: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:09.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:09 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:38:09.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:09 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:38:09.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:09 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:38:10.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:38:10.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:38:10.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:38:10.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:38:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:38:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:38:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:38:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:38:11.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:11 vm01 ceph-mon[48597]: pgmap v522: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:11 vm08 ceph-mon[50506]: pgmap v522: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:13.439 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:38:13.439 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:38:13.467 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:38:13.467 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:38:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:13 vm08 ceph-mon[50506]: pgmap v523: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:38:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:13 vm01 ceph-mon[48597]: pgmap v523: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:38:15.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:15 vm08 ceph-mon[50506]: pgmap v524: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:15.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:15 vm01 ceph-mon[48597]: pgmap v524: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:17.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:17 vm08 ceph-mon[50506]: pgmap v525: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:38:17.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:17 vm01 ceph-mon[48597]: pgmap v525: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:38:18.469 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:38:18.469 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:38:18.495 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:38:18.496 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:38:19.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:19 vm08 ceph-mon[50506]: pgmap v526: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:19.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:19 vm01 ceph-mon[48597]: pgmap v526: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:21.095 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:20 vm08 ceph-mon[50506]: pgmap v527: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:21.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:20 vm01 ceph-mon[48597]: pgmap v527: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:23.497 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:38:23.498 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:38:23.526 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:38:23.526 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:38:23.583 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:23 vm01 ceph-mon[48597]: pgmap v528: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:38:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:23 vm08 ceph-mon[50506]: pgmap v528: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:38:25.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:25 vm08 ceph-mon[50506]: pgmap v529: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:25.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:25 vm01 ceph-mon[48597]: pgmap v529: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:27 vm08 ceph-mon[50506]: pgmap v530: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:38:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:27 vm01 ceph-mon[48597]: pgmap v530: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:38:28.527 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:38:28.528 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:38:28.555 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:38:28.555 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:38:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:29 vm08 ceph-mon[50506]: pgmap v531: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:29.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:29 vm01 ceph-mon[48597]: pgmap v531: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:31.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:30 vm08 ceph-mon[50506]: pgmap v532: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:31.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:30 vm01 ceph-mon[48597]: pgmap v532: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:33.557 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:38:33.557 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:38:33.583 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:33 vm01 ceph-mon[48597]: pgmap v533: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:38:33.585 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:38:33.586 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:38:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:33 vm08 ceph-mon[50506]: pgmap v533: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:38:35.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:35 vm08 ceph-mon[50506]: pgmap v534: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:35.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:35 vm01 ceph-mon[48597]: pgmap v534: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:37 vm08 ceph-mon[50506]: pgmap v535: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:38:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:37 vm01 ceph-mon[48597]: pgmap v535: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:38:38.588 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:38:38.588 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:38:38.614 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:38:38.614 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:38:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:39 vm08 ceph-mon[50506]: pgmap v536: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:39.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:39 vm01 ceph-mon[48597]: pgmap v536: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:41.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:40 vm08 ceph-mon[50506]: pgmap v537: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:41.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:40 vm01 ceph-mon[48597]: pgmap v537: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:43.584 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:43 vm01 ceph-mon[48597]: pgmap v538: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:38:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:43 vm08 ceph-mon[50506]: pgmap v538: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:38:43.615 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:38:43.616 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:38:43.642 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:38:43.642 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:38:45.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:45 vm08 ceph-mon[50506]: pgmap v539: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:45.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:45 vm01 ceph-mon[48597]: pgmap v539: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:47.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:47 vm08 ceph-mon[50506]: pgmap v540: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:38:47.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:47 vm01 ceph-mon[48597]: pgmap v540: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:38:48.644 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:38:48.644 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:38:48.669 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:38:48.670 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:38:49.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:49 vm08 ceph-mon[50506]: pgmap v541: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:49.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:49 vm01 ceph-mon[48597]: pgmap v541: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:51.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:50 vm08 ceph-mon[50506]: pgmap v542: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:51.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:50 vm01 ceph-mon[48597]: pgmap v542: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:53.584 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:53 vm01 ceph-mon[48597]: pgmap v543: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:38:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:53 vm08 ceph-mon[50506]: pgmap v543: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:38:53.671 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:38:53.672 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:38:53.699 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:38:53.700 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:38:55.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:55 vm08 ceph-mon[50506]: pgmap v544: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:55 vm01 ceph-mon[48597]: pgmap v544: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:57 vm08 ceph-mon[50506]: pgmap v545: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:38:57.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:57 vm01 ceph-mon[48597]: pgmap v545: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:38:58.701 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:38:58.702 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:38:58.731 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:38:58.732 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:38:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:38:59 vm01 ceph-mon[48597]: pgmap v546: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:38:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:38:59 vm08 ceph-mon[50506]: pgmap v546: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:01.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:00 vm08 ceph-mon[50506]: pgmap v547: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:01.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:00 vm01 ceph-mon[48597]: pgmap v547: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:03.585 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:03 vm01 ceph-mon[48597]: pgmap v548: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:39:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:03 vm08 ceph-mon[50506]: pgmap v548: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:39:03.733 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:39:03.733 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:39:03.760 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:39:03.760 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:39:05.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:05 vm08 ceph-mon[50506]: pgmap v549: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:05.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:05 vm01 ceph-mon[48597]: pgmap v549: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:07 vm08 ceph-mon[50506]: pgmap v550: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:39:07.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:07 vm01 ceph-mon[48597]: pgmap v550: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:39:08.762 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:39:08.762 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:39:08.789 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:39:08.789 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:39:09.570 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:09 vm01 ceph-mon[48597]: pgmap v551: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:09 vm08 ceph-mon[50506]: pgmap v551: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:10.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:39:10.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:39:10.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:39:10.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:39:10.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:39:10.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:39:10.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:39:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:39:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:39:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:39:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:39:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:39:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:39:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:39:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:11 vm08 ceph-mon[50506]: pgmap v552: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:11.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:11 vm01 ceph-mon[48597]: pgmap v552: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:13.586 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:13 vm01 ceph-mon[48597]: pgmap v553: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:39:13.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:13 vm08 ceph-mon[50506]: pgmap v553: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:39:13.790 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:39:13.791 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:39:13.912 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:39:13.912 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:39:15.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:15 vm08 ceph-mon[50506]: pgmap v554: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:15.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:15 vm01 ceph-mon[48597]: pgmap v554: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:17 vm08 ceph-mon[50506]: pgmap v555: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:39:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:17 vm01 ceph-mon[48597]: pgmap v555: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:39:18.913 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:39:18.914 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:39:18.941 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:39:18.941 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:39:19.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:19 vm08 ceph-mon[50506]: pgmap v556: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:19 vm01 ceph-mon[48597]: pgmap v556: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:21.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:20 vm08 ceph-mon[50506]: pgmap v557: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:21.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:20 vm01 ceph-mon[48597]: pgmap v557: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:23.587 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:23 vm01 ceph-mon[48597]: pgmap v558: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:39:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:23 vm08 ceph-mon[50506]: pgmap v558: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:39:23.942 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:39:23.943 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:39:24.183 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:39:24.184 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:39:25.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:25 vm01 ceph-mon[48597]: pgmap v559: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:25.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:25 vm08 ceph-mon[50506]: pgmap v559: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:27.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:27 vm01 ceph-mon[48597]: pgmap v560: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:39:27.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:27 vm08 ceph-mon[50506]: pgmap v560: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:39:29.186 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:39:29.186 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:39:29.212 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:39:29.213 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:39:29.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:29 vm01 ceph-mon[48597]: pgmap v561: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:29 vm08 ceph-mon[50506]: pgmap v561: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:31.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:30 vm08 ceph-mon[50506]: pgmap v562: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:31.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:30 vm01 ceph-mon[48597]: pgmap v562: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:33.587 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:33 vm01 ceph-mon[48597]: pgmap v563: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:39:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:33 vm08 ceph-mon[50506]: pgmap v563: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:39:34.214 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:39:34.214 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:39:34.338 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:39:34.339 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:39:35.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:35 vm08 ceph-mon[50506]: pgmap v564: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:35.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:35 vm01 ceph-mon[48597]: pgmap v564: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:37 vm08 ceph-mon[50506]: pgmap v565: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:39:37.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:37 vm01 ceph-mon[48597]: pgmap v565: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:39:39.340 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:39:39.341 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:39:39.367 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:39:39.368 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:39:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:39 vm08 ceph-mon[50506]: pgmap v566: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:39.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:39 vm01 ceph-mon[48597]: pgmap v566: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:41 vm01 ceph-mon[48597]: pgmap v567: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:41.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:41 vm08 ceph-mon[50506]: pgmap v567: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:43.588 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:43 vm01 ceph-mon[48597]: pgmap v568: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:39:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:43 vm08 ceph-mon[50506]: pgmap v568: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:39:44.369 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:39:44.370 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:39:44.447 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:39:44.447 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:39:45.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:45 vm08 ceph-mon[50506]: pgmap v569: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:45.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:45 vm01 ceph-mon[48597]: pgmap v569: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:47.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:47 vm08 ceph-mon[50506]: pgmap v570: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:39:47.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:47 vm01 ceph-mon[48597]: pgmap v570: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:39:49.449 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:39:49.450 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:39:49.476 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:39:49.477 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:39:49.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:49 vm08 ceph-mon[50506]: pgmap v571: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:49.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:49 vm01 ceph-mon[48597]: pgmap v571: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:51.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:50 vm01 ceph-mon[48597]: pgmap v572: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:51.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:50 vm08 ceph-mon[50506]: pgmap v572: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:53.589 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:53 vm01 ceph-mon[48597]: pgmap v573: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:39:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:53 vm08 ceph-mon[50506]: pgmap v573: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:39:54.478 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:39:54.479 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:39:54.505 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:39:54.506 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:39:55.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:55 vm08 ceph-mon[50506]: pgmap v574: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:55 vm01 ceph-mon[48597]: pgmap v574: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:57 vm08 ceph-mon[50506]: pgmap v575: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:39:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:57 vm01 ceph-mon[48597]: pgmap v575: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:39:59.508 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:39:59.508 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:39:59.533 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:39:59.533 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:39:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:39:59 vm08 ceph-mon[50506]: pgmap v576: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:39:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:39:59 vm01 ceph-mon[48597]: pgmap v576: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:01.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:00 vm08 ceph-mon[50506]: overall HEALTH_OK 2026-03-10T10:40:01.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:00 vm08 ceph-mon[50506]: pgmap v577: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:00 vm01 ceph-mon[48597]: overall HEALTH_OK 2026-03-10T10:40:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:00 vm01 ceph-mon[48597]: pgmap v577: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:03.590 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:03 vm01 ceph-mon[48597]: pgmap v578: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:40:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:03 vm08 ceph-mon[50506]: pgmap v578: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:40:04.535 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:40:04.535 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:40:04.562 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:40:04.562 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:40:05.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:05 vm08 ceph-mon[50506]: pgmap v579: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:05.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:05 vm01 ceph-mon[48597]: pgmap v579: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:07 vm08 ceph-mon[50506]: pgmap v580: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:40:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:07 vm01 ceph-mon[48597]: pgmap v580: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:40:09.563 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:40:09.564 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:40:09.591 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:40:09.591 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:40:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:09 vm08 ceph-mon[50506]: pgmap v581: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:09.612 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:09 vm01 ceph-mon[48597]: pgmap v581: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:10.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:40:10.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:40:10.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:40:10.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:40:10.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:40:10.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:40:10.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:40:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:40:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:40:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:40:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:40:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:40:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:40:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:40:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:11 vm08 ceph-mon[50506]: pgmap v582: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:11.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:11 vm01 ceph-mon[48597]: pgmap v582: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:13 vm08 ceph-mon[50506]: pgmap v583: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:40:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:13 vm01 ceph-mon[48597]: pgmap v583: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:40:14.593 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:40:14.593 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:40:14.621 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:40:14.621 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:40:15.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:15 vm08 ceph-mon[50506]: pgmap v584: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:15.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:15 vm01 ceph-mon[48597]: pgmap v584: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:17.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:17 vm08 ceph-mon[50506]: pgmap v585: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:40:17.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:17 vm01 ceph-mon[48597]: pgmap v585: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:40:19.623 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:40:19.623 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:40:19.650 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:40:19.651 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:40:19.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:19 vm08 ceph-mon[50506]: pgmap v586: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:19.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:19 vm01 ceph-mon[48597]: pgmap v586: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:20 vm08 ceph-mon[50506]: pgmap v587: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:20 vm01 ceph-mon[48597]: pgmap v587: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:23.592 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:23 vm01 ceph-mon[48597]: pgmap v588: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:40:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:23 vm08 ceph-mon[50506]: pgmap v588: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:40:24.652 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:40:24.652 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:40:24.707 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:40:24.708 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:40:25.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:25 vm01 ceph-mon[48597]: pgmap v589: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:25.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:25 vm08 ceph-mon[50506]: pgmap v589: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:27 vm01 ceph-mon[48597]: pgmap v590: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:40:27.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:27 vm08 ceph-mon[50506]: pgmap v590: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:40:29.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:29 vm01 ceph-mon[48597]: pgmap v591: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:29.709 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:40:29.709 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:40:29.734 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:40:29.735 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:40:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:29 vm08 ceph-mon[50506]: pgmap v591: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:31.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:30 vm08 ceph-mon[50506]: pgmap v592: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:31.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:30 vm01 ceph-mon[48597]: pgmap v592: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:33.592 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:33 vm01 ceph-mon[48597]: pgmap v593: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:40:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:33 vm08 ceph-mon[50506]: pgmap v593: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:40:34.736 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:40:34.737 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:40:34.765 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:40:34.766 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:40:35.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:35 vm08 ceph-mon[50506]: pgmap v594: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:35.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:35 vm01 ceph-mon[48597]: pgmap v594: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:37 vm08 ceph-mon[50506]: pgmap v595: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:40:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:37 vm01 ceph-mon[48597]: pgmap v595: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:40:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:39 vm08 ceph-mon[50506]: pgmap v596: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:39.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:39 vm01 ceph-mon[48597]: pgmap v596: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:39.767 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:40:39.783 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:40:39.814 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:40:39.815 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:40:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:41 vm01 ceph-mon[48597]: pgmap v597: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:41.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:41 vm08 ceph-mon[50506]: pgmap v597: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:43.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:43 vm08 ceph-mon[50506]: pgmap v598: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:40:43.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:43 vm01 ceph-mon[48597]: pgmap v598: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:40:44.816 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:40:44.817 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:40:44.843 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:40:44.844 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:40:45.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:44 vm08 ceph-mon[50506]: pgmap v599: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:45.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:44 vm01 ceph-mon[48597]: pgmap v599: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:47.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:47 vm08 ceph-mon[50506]: pgmap v600: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:40:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:47 vm01 ceph-mon[48597]: pgmap v600: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:40:49.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:49 vm08 ceph-mon[50506]: pgmap v601: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:49.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:49 vm01 ceph-mon[48597]: pgmap v601: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:49.845 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:40:49.846 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:40:49.871 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:40:49.872 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:40:51.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:50 vm08 ceph-mon[50506]: pgmap v602: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:50 vm01 ceph-mon[48597]: pgmap v602: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:53.594 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:53 vm01 ceph-mon[48597]: pgmap v603: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:40:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:53 vm08 ceph-mon[50506]: pgmap v603: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:40:54.873 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:40:54.873 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:40:54.899 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:40:54.900 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:40:55.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:55 vm08 ceph-mon[50506]: pgmap v604: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:55 vm01 ceph-mon[48597]: pgmap v604: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:57 vm08 ceph-mon[50506]: pgmap v605: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:40:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:57 vm01 ceph-mon[48597]: pgmap v605: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:40:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:40:59 vm01 ceph-mon[48597]: pgmap v606: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:40:59 vm08 ceph-mon[50506]: pgmap v606: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:40:59.901 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:40:59.901 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:40:59.926 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:40:59.927 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:41:01.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:00 vm08 ceph-mon[50506]: pgmap v607: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:00 vm01 ceph-mon[48597]: pgmap v607: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:03 vm08 ceph-mon[50506]: pgmap v608: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:41:03.596 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:03 vm01 ceph-mon[48597]: pgmap v608: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:41:04.928 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:41:04.929 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:41:04.954 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:41:04.955 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:41:05.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:05 vm08 ceph-mon[50506]: pgmap v609: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:05.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:05 vm01 ceph-mon[48597]: pgmap v609: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:07.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:07 vm08 ceph-mon[50506]: pgmap v610: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:41:07.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:07 vm01 ceph-mon[48597]: pgmap v610: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:41:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:09 vm08 ceph-mon[50506]: pgmap v611: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:09.866 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:09 vm01 ceph-mon[48597]: pgmap v611: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:09.956 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:41:09.956 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:41:09.982 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:41:09.983 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:41:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:41:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:41:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:41:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:41:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:41:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:41:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:41:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:41:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:41:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:41:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:41:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:41:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:41:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:41:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:11 vm08 ceph-mon[50506]: pgmap v612: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:11 vm01 ceph-mon[48597]: pgmap v612: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:13 vm08 ceph-mon[50506]: pgmap v613: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:41:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:13 vm01 ceph-mon[48597]: pgmap v613: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:41:14.984 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:41:14.985 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:41:15.011 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:41:15.011 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:41:15.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:15 vm08 ceph-mon[50506]: pgmap v614: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:15.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:15 vm01 ceph-mon[48597]: pgmap v614: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:17.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:17 vm08 ceph-mon[50506]: pgmap v615: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:41:17.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:17 vm01 ceph-mon[48597]: pgmap v615: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:41:19.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:19 vm08 ceph-mon[50506]: pgmap v616: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:19.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:19 vm01 ceph-mon[48597]: pgmap v616: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:20.012 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:41:20.013 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:41:20.038 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:41:20.038 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:41:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:20 vm08 ceph-mon[50506]: pgmap v617: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:20 vm01 ceph-mon[48597]: pgmap v617: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:23 vm08 ceph-mon[50506]: pgmap v618: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:41:23.598 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:23 vm01 ceph-mon[48597]: pgmap v618: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:41:25.040 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:41:25.040 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:41:25.067 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:41:25.067 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:41:25.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:25 vm08 ceph-mon[50506]: pgmap v619: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:25.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:25 vm01 ceph-mon[48597]: pgmap v619: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:27 vm08 ceph-mon[50506]: pgmap v620: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:41:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:27 vm01 ceph-mon[48597]: pgmap v620: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:41:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:29 vm08 ceph-mon[50506]: pgmap v621: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:29.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:29 vm01 ceph-mon[48597]: pgmap v621: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:30.068 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:41:30.069 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:41:30.093 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:41:30.094 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:41:31.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:30 vm08 ceph-mon[50506]: pgmap v622: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:31.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:30 vm01 ceph-mon[48597]: pgmap v622: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:33 vm08 ceph-mon[50506]: pgmap v623: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:41:33.599 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:33 vm01 ceph-mon[48597]: pgmap v623: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:41:35.095 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:41:35.096 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:41:35.121 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:41:35.121 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:41:35.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:35 vm08 ceph-mon[50506]: pgmap v624: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:35.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:35 vm01 ceph-mon[48597]: pgmap v624: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:37 vm08 ceph-mon[50506]: pgmap v625: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:41:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:37 vm01 ceph-mon[48597]: pgmap v625: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:41:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:39 vm08 ceph-mon[50506]: pgmap v626: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:39.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:39 vm01 ceph-mon[48597]: pgmap v626: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:40.123 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:41:40.123 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:41:40.150 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:41:40.150 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:41:41.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:40 vm08 ceph-mon[50506]: pgmap v627: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:40 vm01 ceph-mon[48597]: pgmap v627: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:43 vm08 ceph-mon[50506]: pgmap v628: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:41:43.600 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:43 vm01 ceph-mon[48597]: pgmap v628: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:41:45.152 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:41:45.152 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:41:45.180 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:41:45.180 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:41:45.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:45 vm08 ceph-mon[50506]: pgmap v629: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:45.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:45 vm01 ceph-mon[48597]: pgmap v629: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:47.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:47 vm08 ceph-mon[50506]: pgmap v630: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:41:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:47 vm01 ceph-mon[48597]: pgmap v630: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:41:49.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:49 vm01 ceph-mon[48597]: pgmap v631: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:49.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:49 vm08 ceph-mon[50506]: pgmap v631: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:50.181 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:41:50.182 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:41:50.207 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:41:50.208 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:41:51.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:50 vm08 ceph-mon[50506]: pgmap v632: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:50 vm01 ceph-mon[48597]: pgmap v632: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:53 vm08 ceph-mon[50506]: pgmap v633: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:41:53.601 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:53 vm01 ceph-mon[48597]: pgmap v633: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:41:55.209 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:41:55.210 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:41:55.237 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:41:55.237 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:41:55.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:55 vm08 ceph-mon[50506]: pgmap v634: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:55 vm01 ceph-mon[48597]: pgmap v634: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:57 vm08 ceph-mon[50506]: pgmap v635: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:41:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:57 vm01 ceph-mon[48597]: pgmap v635: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:41:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:41:59 vm08 ceph-mon[50506]: pgmap v636: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:41:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:41:59 vm01 ceph-mon[48597]: pgmap v636: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:00.238 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:42:00.239 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:42:00.263 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:42:00.264 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:42:01.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:00 vm08 ceph-mon[50506]: pgmap v637: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:00 vm01 ceph-mon[48597]: pgmap v637: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:03 vm08 ceph-mon[50506]: pgmap v638: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:42:03.602 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:03 vm01 ceph-mon[48597]: pgmap v638: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:42:05.265 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:42:05.266 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:42:05.324 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:42:05.325 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:42:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:05 vm08 ceph-mon[50506]: pgmap v639: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:05 vm01 ceph-mon[48597]: pgmap v639: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:07.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:07 vm08 ceph-mon[50506]: pgmap v640: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:42:07.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:07 vm01 ceph-mon[48597]: pgmap v640: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:42:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:09 vm08 ceph-mon[50506]: pgmap v641: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:09 vm01 ceph-mon[48597]: pgmap v641: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:10.326 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:42:10.327 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:42:10.354 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:42:10.355 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:42:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:42:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:42:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:42:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:42:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:42:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:10 vm08 ceph-mon[50506]: pgmap v642: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:42:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:42:11.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:10 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:42:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:42:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:42:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:42:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:42:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:42:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:10 vm01 ceph-mon[48597]: pgmap v642: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:42:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:42:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:10 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:42:13.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:13 vm08 ceph-mon[50506]: pgmap v643: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:42:13.603 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:13 vm01 ceph-mon[48597]: pgmap v643: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:42:15.358 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:42:15.359 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:42:15.385 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:42:15.385 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:42:15.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:15 vm08 ceph-mon[50506]: pgmap v644: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:15.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:15 vm01 ceph-mon[48597]: pgmap v644: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:17 vm08 ceph-mon[50506]: pgmap v645: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:42:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:17 vm01 ceph-mon[48597]: pgmap v645: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:42:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:19 vm01 ceph-mon[48597]: pgmap v646: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:19.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:19 vm08 ceph-mon[50506]: pgmap v646: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:20.387 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:42:20.387 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:42:20.412 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:42:20.413 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:42:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:20 vm08 ceph-mon[50506]: pgmap v647: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:20 vm01 ceph-mon[48597]: pgmap v647: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:23 vm08 ceph-mon[50506]: pgmap v648: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:42:23.604 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:23 vm01 ceph-mon[48597]: pgmap v648: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:42:25.414 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:42:25.415 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:42:25.442 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:42:25.443 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:42:25.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:25 vm08 ceph-mon[50506]: pgmap v649: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:25.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:25 vm01 ceph-mon[48597]: pgmap v649: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:27 vm01 ceph-mon[48597]: pgmap v650: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:42:27.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:27 vm08 ceph-mon[50506]: pgmap v650: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:42:29.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:29 vm01 ceph-mon[48597]: pgmap v651: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:29 vm08 ceph-mon[50506]: pgmap v651: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:30.444 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:42:30.445 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:42:30.470 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:42:30.471 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:42:31.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:30 vm08 ceph-mon[50506]: pgmap v652: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:31.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:30 vm01 ceph-mon[48597]: pgmap v652: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:33 vm08 ceph-mon[50506]: pgmap v653: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:42:33.604 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:33 vm01 ceph-mon[48597]: pgmap v653: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:42:35.472 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:42:35.472 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:42:35.498 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:42:35.498 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:42:35.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:35 vm08 ceph-mon[50506]: pgmap v654: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:35.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:35 vm01 ceph-mon[48597]: pgmap v654: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:37 vm08 ceph-mon[50506]: pgmap v655: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:42:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:37 vm01 ceph-mon[48597]: pgmap v655: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:42:39.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:39 vm01 ceph-mon[48597]: pgmap v656: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:39 vm08 ceph-mon[50506]: pgmap v656: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:40.499 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:42:40.500 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:42:40.525 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:42:40.526 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:42:41.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:40 vm08 ceph-mon[50506]: pgmap v657: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:40 vm01 ceph-mon[48597]: pgmap v657: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:43 vm08 ceph-mon[50506]: pgmap v658: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:42:43.604 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:43 vm01 ceph-mon[48597]: pgmap v658: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:42:45.527 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:42:45.528 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:42:45.553 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:42:45.554 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:42:45.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:45 vm08 ceph-mon[50506]: pgmap v659: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:45.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:45 vm01 ceph-mon[48597]: pgmap v659: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:47 vm01 ceph-mon[48597]: pgmap v660: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:42:47.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:47 vm08 ceph-mon[50506]: pgmap v660: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:42:49.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:49 vm01 ceph-mon[48597]: pgmap v661: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:49.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:49 vm08 ceph-mon[50506]: pgmap v661: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:50.555 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:42:50.556 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:42:50.582 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:42:50.582 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:42:51.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:51 vm08 ceph-mon[50506]: pgmap v662: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:50 vm01 ceph-mon[48597]: pgmap v662: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:53 vm08 ceph-mon[50506]: pgmap v663: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:53.605 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:53 vm01 ceph-mon[48597]: pgmap v663: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:55.584 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:42:55.584 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:42:55.611 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:42:55.612 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:42:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:55 vm01 ceph-mon[48597]: pgmap v664: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T10:42:55.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:55 vm08 ceph-mon[50506]: pgmap v664: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T10:42:57.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:57 vm08 ceph-mon[50506]: pgmap v665: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:57.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:57 vm01 ceph-mon[48597]: pgmap v665: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:42:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:42:59 vm08 ceph-mon[50506]: pgmap v666: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T10:42:59.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:42:59 vm01 ceph-mon[48597]: pgmap v666: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T10:43:00.613 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:43:00.614 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:43:00.640 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:43:00.640 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:43:01.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:00 vm08 ceph-mon[50506]: pgmap v667: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T10:43:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:00 vm01 ceph-mon[48597]: pgmap v667: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T10:43:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:03 vm08 ceph-mon[50506]: pgmap v668: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:03.607 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:03 vm01 ceph-mon[48597]: pgmap v668: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:05.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:05 vm01 ceph-mon[48597]: pgmap v669: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:05.641 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:43:05.642 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:43:05.710 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:43:05.710 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:43:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:05 vm08 ceph-mon[50506]: pgmap v669: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:07.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:07 vm08 ceph-mon[50506]: pgmap v670: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:43:07.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:07 vm01 ceph-mon[48597]: pgmap v670: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:43:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:09 vm08 ceph-mon[50506]: pgmap v671: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:09 vm01 ceph-mon[48597]: pgmap v671: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:10.712 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:43:10.712 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:43:10.857 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:43:10.857 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:43:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:11 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:43:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:11 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:43:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:11 vm08 ceph-mon[50506]: pgmap v672: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:11 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:43:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:11 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:43:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:11 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:43:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:11 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:43:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:11 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:43:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:43:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:43:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:11 vm01 ceph-mon[48597]: pgmap v672: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:43:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:43:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:43:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:43:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:43:12.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:12 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:43:12.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:12 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:43:12.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:12 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:43:12.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:12 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:43:13.496 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:13 vm08 ceph-mon[50506]: pgmap v673: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:43:13.607 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:13 vm01 ceph-mon[48597]: pgmap v673: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:43:15.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:15 vm08 ceph-mon[50506]: pgmap v674: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:15.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:15 vm01 ceph-mon[48597]: pgmap v674: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:15.858 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:43:15.859 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:43:15.884 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:43:15.884 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:43:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:17 vm01 ceph-mon[48597]: pgmap v675: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:43:17.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:17 vm08 ceph-mon[50506]: pgmap v675: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:43:19.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:19 vm08 ceph-mon[50506]: pgmap v676: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:19.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:19 vm01 ceph-mon[48597]: pgmap v676: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:20.886 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:43:20.886 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:43:20.917 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:43:20.917 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:43:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:21 vm08 ceph-mon[50506]: pgmap v677: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:21 vm01 ceph-mon[48597]: pgmap v677: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:23.608 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:23 vm01 ceph-mon[48597]: pgmap v678: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:43:23.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:23 vm08 ceph-mon[50506]: pgmap v678: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:43:25.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:25 vm01 ceph-mon[48597]: pgmap v679: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:25.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:25 vm08 ceph-mon[50506]: pgmap v679: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:25.919 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:43:25.919 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:43:25.946 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:43:25.947 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:43:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:27 vm01 ceph-mon[48597]: pgmap v680: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:43:27.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:27 vm08 ceph-mon[50506]: pgmap v680: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:43:29.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:29 vm01 ceph-mon[48597]: pgmap v681: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:29 vm08 ceph-mon[50506]: pgmap v681: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:30.948 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:43:30.948 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:43:30.975 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:43:30.975 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:43:31.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:31 vm08 ceph-mon[50506]: pgmap v682: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:31.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:31 vm01 ceph-mon[48597]: pgmap v682: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:33.609 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:33 vm01 ceph-mon[48597]: pgmap v683: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:43:33.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:33 vm08 ceph-mon[50506]: pgmap v683: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:43:35.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:35 vm01 ceph-mon[48597]: pgmap v684: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:35.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:35 vm08 ceph-mon[50506]: pgmap v684: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:35.976 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:43:35.977 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:43:36.003 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:43:36.004 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:43:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:37 vm01 ceph-mon[48597]: pgmap v685: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:43:37.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:37 vm08 ceph-mon[50506]: pgmap v685: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:43:39.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:39 vm01 ceph-mon[48597]: pgmap v686: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:39 vm08 ceph-mon[50506]: pgmap v686: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:41.005 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:43:41.005 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:43:41.031 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:43:41.032 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:43:41.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:41 vm08 ceph-mon[50506]: pgmap v687: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:41 vm01 ceph-mon[48597]: pgmap v687: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:43.609 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:43 vm01 ceph-mon[48597]: pgmap v688: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:43:43.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:43 vm08 ceph-mon[50506]: pgmap v688: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:43:45.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:45 vm01 ceph-mon[48597]: pgmap v689: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:45.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:45 vm08 ceph-mon[50506]: pgmap v689: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:46.033 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:43:46.033 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:43:46.061 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:43:46.062 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:43:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:47 vm01 ceph-mon[48597]: pgmap v690: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:43:47.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:47 vm08 ceph-mon[50506]: pgmap v690: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:43:50.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:49 vm08 ceph-mon[50506]: pgmap v691: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:50.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:49 vm01 ceph-mon[48597]: pgmap v691: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:51.063 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:43:51.063 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:43:51.090 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:43:51.090 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:43:51.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:51 vm08 ceph-mon[50506]: pgmap v692: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:51 vm01 ceph-mon[48597]: pgmap v692: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:53.610 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:53 vm01 ceph-mon[48597]: pgmap v693: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:43:53.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:53 vm08 ceph-mon[50506]: pgmap v693: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:43:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:55 vm01 ceph-mon[48597]: pgmap v694: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:55.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:55 vm08 ceph-mon[50506]: pgmap v694: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:56.091 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:43:56.092 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:43:56.117 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:43:56.117 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:43:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:57 vm01 ceph-mon[48597]: pgmap v695: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:43:57.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:57 vm08 ceph-mon[50506]: pgmap v695: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:43:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:43:59 vm01 ceph-mon[48597]: pgmap v696: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:43:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:43:59 vm08 ceph-mon[50506]: pgmap v696: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:01.118 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:44:01.119 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:44:01.144 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:44:01.144 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:44:01.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:01 vm08 ceph-mon[50506]: pgmap v697: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:01 vm01 ceph-mon[48597]: pgmap v697: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:03.611 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:03 vm01 ceph-mon[48597]: pgmap v698: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:44:03.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:03 vm08 ceph-mon[50506]: pgmap v698: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:44:05.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:05 vm01 ceph-mon[48597]: pgmap v699: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:05 vm08 ceph-mon[50506]: pgmap v699: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:06.146 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:44:06.146 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:44:06.172 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:44:06.173 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:44:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:07 vm01 ceph-mon[48597]: pgmap v700: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:44:07.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:07 vm08 ceph-mon[50506]: pgmap v700: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:44:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:09 vm08 ceph-mon[50506]: pgmap v701: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:09 vm01 ceph-mon[48597]: pgmap v701: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:11.174 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:44:11.174 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:44:11.202 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:44:11.202 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:44:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:11 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:44:11.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:11 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:44:11.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:11 vm08 ceph-mon[50506]: pgmap v702: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:11.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:44:11.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:44:11.369 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:11 vm01 ceph-mon[48597]: pgmap v702: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:12.293 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:12 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:44:12.293 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:12 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:44:12.293 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:12 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:44:12.362 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:12 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:44:12.362 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:12 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:44:12.362 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:12 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:44:13.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:13 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:44:13.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:13 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:44:13.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:13 vm08 ceph-mon[50506]: pgmap v703: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:44:13.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:13 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:44:13.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:13 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:44:13.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:13 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:44:13.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:13 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:44:13.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:13 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:44:13.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:13 vm01 ceph-mon[48597]: pgmap v703: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:44:13.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:13 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:44:13.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:13 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:44:13.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:13 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:44:15.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:15 vm08 ceph-mon[50506]: pgmap v704: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:15.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:15 vm01 ceph-mon[48597]: pgmap v704: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:16.203 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:44:16.204 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:44:16.229 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:44:16.229 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:44:17.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:17 vm08 ceph-mon[50506]: pgmap v705: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:44:17.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:17 vm01 ceph-mon[48597]: pgmap v705: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:44:19.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:19 vm08 ceph-mon[50506]: pgmap v706: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:19.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:19 vm01 ceph-mon[48597]: pgmap v706: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:21.231 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:44:21.231 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:44:21.258 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:44:21.259 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:44:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:21 vm08 ceph-mon[50506]: pgmap v707: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:21 vm01 ceph-mon[48597]: pgmap v707: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:23.613 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:23 vm01 ceph-mon[48597]: pgmap v708: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:44:23.731 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:23 vm08 ceph-mon[50506]: pgmap v708: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:44:25.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:25 vm01 ceph-mon[48597]: pgmap v709: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:25.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:25 vm08 ceph-mon[50506]: pgmap v709: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:26.260 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:44:26.261 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:44:26.287 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:44:26.288 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:44:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:27 vm01 ceph-mon[48597]: pgmap v710: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:44:27.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:27 vm08 ceph-mon[50506]: pgmap v710: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:44:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:29 vm08 ceph-mon[50506]: pgmap v711: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:29 vm01 ceph-mon[48597]: pgmap v711: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:31.289 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:44:31.290 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:44:31.316 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:44:31.316 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:44:31.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:31 vm08 ceph-mon[50506]: pgmap v712: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:31.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:31 vm01 ceph-mon[48597]: pgmap v712: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:33.614 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:33 vm01 ceph-mon[48597]: pgmap v713: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:44:33.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:33 vm08 ceph-mon[50506]: pgmap v713: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:44:35.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:35 vm01 ceph-mon[48597]: pgmap v714: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:35.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:35 vm08 ceph-mon[50506]: pgmap v714: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:36.318 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:44:36.318 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:44:36.344 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:44:36.344 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:44:37.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:37 vm08 ceph-mon[50506]: pgmap v715: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:44:37.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:37 vm01 ceph-mon[48597]: pgmap v715: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:44:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:39 vm08 ceph-mon[50506]: pgmap v716: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:39.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:39 vm01 ceph-mon[48597]: pgmap v716: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:41.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:41 vm08 ceph-mon[50506]: pgmap v717: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:41.345 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:44:41.346 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:44:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:41 vm01 ceph-mon[48597]: pgmap v717: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:41.372 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:44:41.372 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:44:43.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:43 vm08 ceph-mon[50506]: pgmap v718: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:44:43.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:43 vm01 ceph-mon[48597]: pgmap v718: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:44:45.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:45 vm08 ceph-mon[50506]: pgmap v719: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:45.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:45 vm01 ceph-mon[48597]: pgmap v719: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:46.373 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:44:46.374 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:44:46.402 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:44:46.403 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:44:47.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:47 vm08 ceph-mon[50506]: pgmap v720: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:44:47.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:47 vm01 ceph-mon[48597]: pgmap v720: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:44:49.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:49 vm08 ceph-mon[50506]: pgmap v721: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:49.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:49 vm01 ceph-mon[48597]: pgmap v721: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:51.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:51 vm08 ceph-mon[50506]: pgmap v722: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:51 vm01 ceph-mon[48597]: pgmap v722: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:51.404 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:44:51.405 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:44:51.431 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:44:51.431 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:44:53.616 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:53 vm01 ceph-mon[48597]: pgmap v723: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:44:53.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:53 vm08 ceph-mon[50506]: pgmap v723: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:44:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:55 vm01 ceph-mon[48597]: pgmap v724: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:55.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:55 vm08 ceph-mon[50506]: pgmap v724: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:56.433 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:44:56.433 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:44:56.459 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:44:56.460 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:44:57.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:57 vm08 ceph-mon[50506]: pgmap v725: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:44:57.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:57 vm01 ceph-mon[48597]: pgmap v725: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:44:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:44:59 vm08 ceph-mon[50506]: pgmap v726: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:44:59.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:44:59 vm01 ceph-mon[48597]: pgmap v726: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:01.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:01 vm08 ceph-mon[50506]: pgmap v727: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:01 vm01 ceph-mon[48597]: pgmap v727: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:01.461 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:45:01.462 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:45:01.487 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:45:01.487 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:45:03.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:03 vm01 ceph-mon[48597]: pgmap v728: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:45:03.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:03 vm08 ceph-mon[50506]: pgmap v728: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:45:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:05 vm08 ceph-mon[50506]: pgmap v729: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:05 vm01 ceph-mon[48597]: pgmap v729: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:06.489 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:45:06.489 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:45:06.516 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:45:06.517 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:45:07.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:07 vm08 ceph-mon[50506]: pgmap v730: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:45:07.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:07 vm01 ceph-mon[48597]: pgmap v730: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:45:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:09 vm08 ceph-mon[50506]: pgmap v731: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:09 vm01 ceph-mon[48597]: pgmap v731: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:11 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:45:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:11 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:45:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:11 vm08 ceph-mon[50506]: pgmap v732: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:45:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:45:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:11 vm01 ceph-mon[48597]: pgmap v732: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:11.518 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:45:11.519 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:45:11.545 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:45:11.545 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:45:13.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:13 vm01 ceph-mon[48597]: pgmap v733: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:45:13.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:13 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:45:13.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:13 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:45:13.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:13 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:45:13.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:13 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:45:13.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:13 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:45:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:13 vm08 ceph-mon[50506]: pgmap v733: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:45:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:13 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:45:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:13 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:45:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:13 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:45:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:13 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:45:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:13 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:45:15.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:15 vm08 ceph-mon[50506]: pgmap v734: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:15.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:15 vm01 ceph-mon[48597]: pgmap v734: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:16.547 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:45:16.547 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:45:16.573 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:45:16.573 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:45:17.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:17 vm08 ceph-mon[50506]: pgmap v735: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:45:17.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:17 vm01 ceph-mon[48597]: pgmap v735: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:45:19.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:19 vm08 ceph-mon[50506]: pgmap v736: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:19.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:19 vm01 ceph-mon[48597]: pgmap v736: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:21 vm08 ceph-mon[50506]: pgmap v737: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:21 vm01 ceph-mon[48597]: pgmap v737: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:21.575 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:45:21.575 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:45:21.601 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:45:21.602 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:45:23.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:23 vm01 ceph-mon[48597]: pgmap v738: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:45:23.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:23 vm08 ceph-mon[50506]: pgmap v738: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:45:25.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:25 vm08 ceph-mon[50506]: pgmap v739: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:25 vm01 ceph-mon[48597]: pgmap v739: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:26.604 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:45:26.604 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:45:26.630 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:45:26.631 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:45:27.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:27 vm08 ceph-mon[50506]: pgmap v740: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:45:27.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:27 vm01 ceph-mon[48597]: pgmap v740: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:45:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:29 vm08 ceph-mon[50506]: pgmap v741: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:29 vm01 ceph-mon[48597]: pgmap v741: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:31.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:31 vm08 ceph-mon[50506]: pgmap v742: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:31.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:31 vm01 ceph-mon[48597]: pgmap v742: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:31.632 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:45:31.633 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:45:31.658 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:45:31.658 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:45:33.620 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:33 vm01 ceph-mon[48597]: pgmap v743: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:45:33.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:33 vm08 ceph-mon[50506]: pgmap v743: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:45:35.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:35 vm08 ceph-mon[50506]: pgmap v744: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:35.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:35 vm01 ceph-mon[48597]: pgmap v744: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:36.660 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:45:36.660 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:45:36.686 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:45:36.686 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:45:37.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:37 vm08 ceph-mon[50506]: pgmap v745: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:45:37.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:37 vm01 ceph-mon[48597]: pgmap v745: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:45:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:39 vm08 ceph-mon[50506]: pgmap v746: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:39.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:39 vm01 ceph-mon[48597]: pgmap v746: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:41.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:41 vm08 ceph-mon[50506]: pgmap v747: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:41 vm01 ceph-mon[48597]: pgmap v747: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:41.688 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:45:41.689 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:45:41.719 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:45:41.720 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:45:43.625 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:43 vm01 ceph-mon[48597]: pgmap v748: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:45:43.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:43 vm08 ceph-mon[50506]: pgmap v748: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:45:45.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:45 vm08 ceph-mon[50506]: pgmap v749: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:45.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:45 vm01 ceph-mon[48597]: pgmap v749: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:46.721 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:45:46.722 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:45:46.796 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:45:46.797 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:45:47.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:47 vm08 ceph-mon[50506]: pgmap v750: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:45:47.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:47 vm01 ceph-mon[48597]: pgmap v750: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:45:49.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:49 vm08 ceph-mon[50506]: pgmap v751: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:49.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:49 vm01 ceph-mon[48597]: pgmap v751: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:51.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:51 vm08 ceph-mon[50506]: pgmap v752: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:51 vm01 ceph-mon[48597]: pgmap v752: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:51.798 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:45:51.799 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:45:51.827 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:45:51.828 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:45:53.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:53 vm08 ceph-mon[50506]: pgmap v753: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:45:53.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:53 vm01 ceph-mon[48597]: pgmap v753: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:45:55.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:55 vm08 ceph-mon[50506]: pgmap v754: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:55.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:55 vm01 ceph-mon[48597]: pgmap v754: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:56.829 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:45:56.830 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:45:56.858 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:45:56.859 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:45:57.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:57 vm08 ceph-mon[50506]: pgmap v755: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:45:57.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:57 vm01 ceph-mon[48597]: pgmap v755: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:45:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:45:59 vm08 ceph-mon[50506]: pgmap v756: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:45:59.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:45:59 vm01 ceph-mon[48597]: pgmap v756: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:01.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:01 vm08 ceph-mon[50506]: pgmap v757: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:01 vm01 ceph-mon[48597]: pgmap v757: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:01.860 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:46:01.861 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:46:01.903 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:46:01.904 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:46:03.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:03 vm08 ceph-mon[50506]: pgmap v758: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:46:03.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:03 vm01 ceph-mon[48597]: pgmap v758: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:46:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:05 vm08 ceph-mon[50506]: pgmap v759: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:05 vm01 ceph-mon[48597]: pgmap v759: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:06.905 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:46:06.906 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:46:06.931 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:46:06.932 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:46:07.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:07 vm08 ceph-mon[50506]: pgmap v760: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:46:07.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:07 vm01 ceph-mon[48597]: pgmap v760: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:46:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:09 vm08 ceph-mon[50506]: pgmap v761: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:09 vm01 ceph-mon[48597]: pgmap v761: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:11 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:46:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:11 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:46:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:11 vm08 ceph-mon[50506]: pgmap v762: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:46:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:11 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:46:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:11 vm01 ceph-mon[48597]: pgmap v762: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:11.933 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:46:11.934 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:46:11.960 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:46:11.960 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:46:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:13 vm08 ceph-mon[50506]: pgmap v763: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:46:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:13 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:46:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:13 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:46:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:13 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:46:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:13 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:46:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:13 vm08 ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:46:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:13 vm01 ceph-mon[48597]: pgmap v763: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:46:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:13 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:46:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:13 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:46:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:13 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:46:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:13 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:46:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:13 vm01 ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:46:15.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:15 vm08 ceph-mon[50506]: pgmap v764: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:15.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:15 vm01 ceph-mon[48597]: pgmap v764: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:16.961 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:46:16.962 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:46:16.989 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:46:16.989 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:46:17.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:17 vm08 ceph-mon[50506]: pgmap v765: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:46:17.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:17 vm01 ceph-mon[48597]: pgmap v765: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:46:19.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:19 vm08 ceph-mon[50506]: pgmap v766: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:19.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:19 vm01 ceph-mon[48597]: pgmap v766: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:21 vm08 ceph-mon[50506]: pgmap v767: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:21 vm01 ceph-mon[48597]: pgmap v767: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:21.990 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:46:21.991 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:46:22.016 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:46:22.016 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:46:23.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:23 vm08 ceph-mon[50506]: pgmap v768: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:46:23.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:23 vm01 ceph-mon[48597]: pgmap v768: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:46:25.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:25 vm08 ceph-mon[50506]: pgmap v769: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:25 vm01 ceph-mon[48597]: pgmap v769: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:27.018 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:46:27.019 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:46:27.045 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:46:27.046 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:46:27.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:27 vm08 ceph-mon[50506]: pgmap v770: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:46:27.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:27 vm01 ceph-mon[48597]: pgmap v770: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:46:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:29 vm08 ceph-mon[50506]: pgmap v771: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:29 vm01 ceph-mon[48597]: pgmap v771: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:31.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:31 vm01 ceph-mon[48597]: pgmap v772: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:31.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:31 vm08 ceph-mon[50506]: pgmap v772: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:32.047 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:46:32.047 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:46:32.072 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:46:32.072 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:46:33.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:33 vm08 ceph-mon[50506]: pgmap v773: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:46:33.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:33 vm01 ceph-mon[48597]: pgmap v773: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:46:35.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:35 vm08 ceph-mon[50506]: pgmap v774: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:35.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:35 vm01 ceph-mon[48597]: pgmap v774: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:37.074 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:46:37.074 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:46:37.101 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:46:37.101 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:46:37.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:37 vm08 ceph-mon[50506]: pgmap v775: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:46:37.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:37 vm01 ceph-mon[48597]: pgmap v775: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:46:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:39 vm08 ceph-mon[50506]: pgmap v776: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:39.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:39 vm01 ceph-mon[48597]: pgmap v776: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:41 vm01 ceph-mon[48597]: pgmap v777: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:41.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:41 vm08 ceph-mon[50506]: pgmap v777: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:42.103 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:46:42.103 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01:/fake /mnt/foo -o sync 2026-03-10T10:46:42.130 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:46:42.131 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:46:43.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:43 vm01.local ceph-mon[48597]: pgmap v778: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:46:43.656 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:43 vm08 ceph-mon[50506]: pgmap v778: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:46:45.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:45 vm08 ceph-mon[50506]: pgmap v779: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:45.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:45 vm01.local ceph-mon[48597]: pgmap v779: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:47.133 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:46:47.134 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:46:47.161 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:46:47.161 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:46:47.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:47 vm08 ceph-mon[50506]: pgmap v780: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:46:47.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:47 vm01.local ceph-mon[48597]: pgmap v780: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:46:49.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:48 vm08 ceph-mon[50506]: pgmap v781: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:49.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:48 vm01.local ceph-mon[48597]: pgmap v781: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:51 vm01.local ceph-mon[48597]: pgmap v782: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:51.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:51 vm08 ceph-mon[50506]: pgmap v782: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:52.163 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:46:52.164 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:46:52.190 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:46:52.191 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:46:53.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:53 vm01.local ceph-mon[48597]: pgmap v783: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:46:53.796 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:53 vm08 ceph-mon[50506]: pgmap v783: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:46:55.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:55 vm08 ceph-mon[50506]: pgmap v784: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:55.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:55 vm01.local ceph-mon[48597]: pgmap v784: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:57.192 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:46:57.193 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:46:57.220 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:46:57.220 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:46:57.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:57 vm08 ceph-mon[50506]: pgmap v785: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:46:57.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:57 vm01.local ceph-mon[48597]: pgmap v785: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:46:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:46:59 vm08 ceph-mon[50506]: pgmap v786: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:46:59.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:46:59 vm01.local ceph-mon[48597]: pgmap v786: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:01 vm01.local ceph-mon[48597]: pgmap v787: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:01.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:01 vm08 ceph-mon[50506]: pgmap v787: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:02.222 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:47:02.222 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:47:02.249 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:47:02.249 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:47:03.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:03 vm08 ceph-mon[50506]: pgmap v788: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:47:03.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:03 vm01.local ceph-mon[48597]: pgmap v788: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:47:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:05 vm08 ceph-mon[50506]: pgmap v789: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:05 vm01.local ceph-mon[48597]: pgmap v789: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:07.251 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:47:07.251 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:47:07.277 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:47:07.278 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:47:07.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:07 vm08 ceph-mon[50506]: pgmap v790: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:47:07.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:07 vm01.local ceph-mon[48597]: pgmap v790: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:47:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:09 vm08.local ceph-mon[50506]: pgmap v791: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:09 vm01.local ceph-mon[48597]: pgmap v791: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:47:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:47:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:47:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:47:11.641 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:11 vm01.local ceph-mon[48597]: pgmap v792: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:11 vm08.local ceph-mon[50506]: pgmap v792: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:12.280 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:47:12.281 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:47:12.307 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:47:12.307 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:47:13.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:13 vm08.local ceph-mon[50506]: pgmap v793: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:47:13.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:13 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:47:13.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:13 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:47:13.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:13 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:47:13.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:13 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:47:13.595 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:13 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:47:13.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:13 vm01.local ceph-mon[48597]: pgmap v793: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:47:13.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:13 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:47:13.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:13 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:47:13.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:13 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:47:13.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:13 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:47:13.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:13 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:47:15.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:15 vm08.local ceph-mon[50506]: pgmap v794: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:15.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:15 vm01.local ceph-mon[48597]: pgmap v794: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:17.309 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:47:17.310 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:47:17.335 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:47:17.336 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:47:17.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:17 vm08.local ceph-mon[50506]: pgmap v795: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:47:17.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:17 vm01.local ceph-mon[48597]: pgmap v795: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:47:19.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:19 vm01.local ceph-mon[48597]: pgmap v796: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:19 vm08.local ceph-mon[50506]: pgmap v796: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:21 vm01.local ceph-mon[48597]: pgmap v797: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:21.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:21 vm08.local ceph-mon[50506]: pgmap v797: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:22.337 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:47:22.338 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:47:22.364 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:47:22.364 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:47:23.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:23 vm08.local ceph-mon[50506]: pgmap v798: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:47:23.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:23 vm01.local ceph-mon[48597]: pgmap v798: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:47:25.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:25 vm08.local ceph-mon[50506]: pgmap v799: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:25 vm01.local ceph-mon[48597]: pgmap v799: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:27.366 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:47:27.366 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:47:27.391 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:47:27.391 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:47:27.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:27 vm08.local ceph-mon[50506]: pgmap v800: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:47:27.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:27 vm01.local ceph-mon[48597]: pgmap v800: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:47:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:29 vm08.local ceph-mon[50506]: pgmap v801: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:29 vm01.local ceph-mon[48597]: pgmap v801: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:31.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:31 vm01.local ceph-mon[48597]: pgmap v802: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:31.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:31 vm08.local ceph-mon[50506]: pgmap v802: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:32.392 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:47:32.393 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:47:32.513 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:47:32.513 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:47:33.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:33 vm08.local ceph-mon[50506]: pgmap v803: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:47:33.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:33 vm01.local ceph-mon[48597]: pgmap v803: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:47:35.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:35 vm08.local ceph-mon[50506]: pgmap v804: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:35.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:35 vm01.local ceph-mon[48597]: pgmap v804: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:37.515 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:47:37.515 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:47:37.542 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:47:37.542 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:47:37.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:37 vm08.local ceph-mon[50506]: pgmap v805: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:47:37.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:37 vm01.local ceph-mon[48597]: pgmap v805: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:47:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:39 vm08.local ceph-mon[50506]: pgmap v806: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:39.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:39 vm01.local ceph-mon[48597]: pgmap v806: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:41 vm01.local ceph-mon[48597]: pgmap v807: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:41.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:41 vm08.local ceph-mon[50506]: pgmap v807: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:42.544 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:47:42.544 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:47:42.571 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:47:42.572 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:47:43.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:43 vm08.local ceph-mon[50506]: pgmap v808: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:47:43.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:43 vm01.local ceph-mon[48597]: pgmap v808: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:47:45.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:45 vm08.local ceph-mon[50506]: pgmap v809: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:45.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:45 vm01.local ceph-mon[48597]: pgmap v809: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:47.573 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:47:47.574 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:47:47.599 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:47:47.600 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:47:47.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:47 vm08.local ceph-mon[50506]: pgmap v810: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:47:47.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:47 vm01.local ceph-mon[48597]: pgmap v810: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:47:49.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:49 vm08.local ceph-mon[50506]: pgmap v811: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:49.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:49 vm01.local ceph-mon[48597]: pgmap v811: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:51 vm01.local ceph-mon[48597]: pgmap v812: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:51.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:51 vm08.local ceph-mon[50506]: pgmap v812: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:52.601 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:47:52.601 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:47:52.627 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:47:52.627 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:47:53.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:53 vm08.local ceph-mon[50506]: pgmap v813: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:47:53.874 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:53 vm01.local ceph-mon[48597]: pgmap v813: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:47:55.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:55 vm08.local ceph-mon[50506]: pgmap v814: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:55.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:55 vm01.local ceph-mon[48597]: pgmap v814: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:47:57.629 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:47:57.629 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:47:57.659 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:47:57.660 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:47:57.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:57 vm08.local ceph-mon[50506]: pgmap v815: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:47:57.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:57 vm01.local ceph-mon[48597]: pgmap v815: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:48:00.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:47:59 vm08.local ceph-mon[50506]: pgmap v816: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:00.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:47:59 vm01.local ceph-mon[48597]: pgmap v816: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:01 vm01.local ceph-mon[48597]: pgmap v817: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:01.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:01 vm08.local ceph-mon[50506]: pgmap v817: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:02.661 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:48:02.662 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:48:02.688 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:48:02.688 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:48:03.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:03 vm08.local ceph-mon[50506]: pgmap v818: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:48:03.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:03 vm01.local ceph-mon[48597]: pgmap v818: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:48:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:05 vm08.local ceph-mon[50506]: pgmap v819: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:05 vm01.local ceph-mon[48597]: pgmap v819: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:07.690 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:48:07.690 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:48:07.715 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:48:07.715 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:48:07.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:07 vm08.local ceph-mon[50506]: pgmap v820: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:48:07.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:07 vm01.local ceph-mon[48597]: pgmap v820: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:48:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:09 vm08.local ceph-mon[50506]: pgmap v821: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:09 vm01.local ceph-mon[48597]: pgmap v821: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:48:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:48:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:48:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:48:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:11 vm08.local ceph-mon[50506]: pgmap v822: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:11 vm01.local ceph-mon[48597]: pgmap v822: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:12.716 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:48:12.717 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:48:12.741 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:48:12.742 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:48:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:13 vm08.local ceph-mon[50506]: pgmap v823: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:48:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:13 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:48:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:13 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:48:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:13 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:48:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:13 vm01.local ceph-mon[48597]: pgmap v823: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:48:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:13 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:48:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:13 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:48:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:13 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:48:15.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:14 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:48:15.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:14 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:48:15.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:14 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:48:15.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:14 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:48:16.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:15 vm08.local ceph-mon[50506]: pgmap v824: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:15 vm01.local ceph-mon[48597]: pgmap v824: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:17.743 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:48:17.744 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:48:17.770 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:48:17.770 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:48:18.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:17 vm08.local ceph-mon[50506]: pgmap v825: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:48:18.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:17 vm01.local ceph-mon[48597]: pgmap v825: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:48:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:19 vm08.local ceph-mon[50506]: pgmap v826: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:19 vm01.local ceph-mon[48597]: pgmap v826: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:21.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:21 vm08.local ceph-mon[50506]: pgmap v827: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:21.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:21 vm01.local ceph-mon[48597]: pgmap v827: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:22.772 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:48:22.772 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:48:22.799 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:48:22.799 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:48:23.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:23 vm08.local ceph-mon[50506]: pgmap v828: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:48:23.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:23 vm01.local ceph-mon[48597]: pgmap v828: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:48:25.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:25 vm08.local ceph-mon[50506]: pgmap v829: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:25 vm01.local ceph-mon[48597]: pgmap v829: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:27.800 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:48:27.801 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:48:27.826 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:48:27.827 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:48:27.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:27 vm08.local ceph-mon[50506]: pgmap v830: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:48:27.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:27 vm01.local ceph-mon[48597]: pgmap v830: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:48:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:29 vm08.local ceph-mon[50506]: pgmap v831: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:29 vm01.local ceph-mon[48597]: pgmap v831: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:31.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:31 vm01.local ceph-mon[48597]: pgmap v832: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:31.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:31 vm08.local ceph-mon[50506]: pgmap v832: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:32.828 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:48:32.829 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:48:32.855 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:48:32.855 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:48:33.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:33 vm08.local ceph-mon[50506]: pgmap v833: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:48:33.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:33 vm01.local ceph-mon[48597]: pgmap v833: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:48:35.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:35 vm08.local ceph-mon[50506]: pgmap v834: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:35.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:35 vm01.local ceph-mon[48597]: pgmap v834: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:37.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:37 vm08.local ceph-mon[50506]: pgmap v835: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:48:37.856 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:48:37.857 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:48:37.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:37 vm01.local ceph-mon[48597]: pgmap v835: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:48:37.883 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:48:37.884 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:48:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:39 vm08.local ceph-mon[50506]: pgmap v836: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:39.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:39 vm01.local ceph-mon[48597]: pgmap v836: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:41 vm01.local ceph-mon[48597]: pgmap v837: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:41.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:41 vm08.local ceph-mon[50506]: pgmap v837: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:42.885 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:48:42.886 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:48:42.911 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:48:42.911 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:48:43.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:43 vm08.local ceph-mon[50506]: pgmap v838: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:48:43.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:43 vm01.local ceph-mon[48597]: pgmap v838: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:48:45.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:45 vm08.local ceph-mon[50506]: pgmap v839: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:45.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:45 vm01.local ceph-mon[48597]: pgmap v839: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:47.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:47 vm08.local ceph-mon[50506]: pgmap v840: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:48:47.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:47 vm01.local ceph-mon[48597]: pgmap v840: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:48:47.913 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:48:47.913 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:48:47.938 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:48:47.939 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:48:49.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:49 vm08.local ceph-mon[50506]: pgmap v841: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:49.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:49 vm01.local ceph-mon[48597]: pgmap v841: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:51.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:51 vm08.local ceph-mon[50506]: pgmap v842: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:51.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:51 vm01.local ceph-mon[48597]: pgmap v842: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:52.940 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:48:52.941 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:48:52.966 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:48:52.966 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:48:53.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:53 vm08.local ceph-mon[50506]: pgmap v843: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:48:53.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:53 vm01.local ceph-mon[48597]: pgmap v843: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:48:55.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:55 vm08.local ceph-mon[50506]: pgmap v844: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:55.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:55 vm01.local ceph-mon[48597]: pgmap v844: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:57.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:57 vm08.local ceph-mon[50506]: pgmap v845: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:48:57.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:57 vm01.local ceph-mon[48597]: pgmap v845: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:48:57.968 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:48:57.968 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:48:57.996 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:48:57.996 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:48:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:48:59 vm08.local ceph-mon[50506]: pgmap v846: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:48:59.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:48:59 vm01.local ceph-mon[48597]: pgmap v846: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:01.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:01 vm08.local ceph-mon[50506]: pgmap v847: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:01.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:01 vm01.local ceph-mon[48597]: pgmap v847: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:02.998 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:49:02.998 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:49:03.027 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:49:03.027 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:49:03.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:03 vm08.local ceph-mon[50506]: pgmap v848: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:49:03.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:03 vm01.local ceph-mon[48597]: pgmap v848: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:49:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:05 vm08.local ceph-mon[50506]: pgmap v849: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:05 vm01.local ceph-mon[48597]: pgmap v849: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:07.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:07 vm08.local ceph-mon[50506]: pgmap v850: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:49:07.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:07 vm01.local ceph-mon[48597]: pgmap v850: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:49:08.028 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:49:08.029 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:49:08.054 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:49:08.055 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:49:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:09 vm08.local ceph-mon[50506]: pgmap v851: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:09 vm01.local ceph-mon[48597]: pgmap v851: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:49:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:49:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:49:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:49:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:11 vm08.local ceph-mon[50506]: pgmap v852: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:11 vm01.local ceph-mon[48597]: pgmap v852: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:13.056 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:49:13.056 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:49:13.082 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:49:13.083 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:49:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:13 vm08.local ceph-mon[50506]: pgmap v853: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:49:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:13 vm01.local ceph-mon[48597]: pgmap v853: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:49:14.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:14 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:49:14.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:14 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:49:14.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:14 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:49:14.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:14 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:49:14.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:14 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:49:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:14 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:49:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:14 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:49:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:14 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:49:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:14 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:49:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:14 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:49:15.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:15 vm08.local ceph-mon[50506]: pgmap v854: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:15.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:15 vm01.local ceph-mon[48597]: pgmap v854: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:17.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:17 vm08.local ceph-mon[50506]: pgmap v855: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:49:17.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:17 vm01.local ceph-mon[48597]: pgmap v855: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:49:18.085 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:49:18.085 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:49:18.111 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:49:18.111 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:49:19.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:19 vm08.local ceph-mon[50506]: pgmap v856: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:19.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:19 vm01.local ceph-mon[48597]: pgmap v856: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:21.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:21 vm08.local ceph-mon[50506]: pgmap v857: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:21.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:21 vm01.local ceph-mon[48597]: pgmap v857: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:23.113 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:49:23.113 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:49:23.140 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:49:23.141 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:49:23.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:23 vm08.local ceph-mon[50506]: pgmap v858: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:49:23.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:23 vm01.local ceph-mon[48597]: pgmap v858: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:49:25.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:25 vm08.local ceph-mon[50506]: pgmap v859: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:25 vm01.local ceph-mon[48597]: pgmap v859: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:27.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:27 vm08.local ceph-mon[50506]: pgmap v860: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:49:27.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:27 vm01.local ceph-mon[48597]: pgmap v860: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:49:28.142 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:49:28.143 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:49:28.169 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:49:28.169 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:49:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:29 vm08.local ceph-mon[50506]: pgmap v861: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:29 vm01.local ceph-mon[48597]: pgmap v861: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:31.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:31 vm08.local ceph-mon[50506]: pgmap v862: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:31.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:31 vm01.local ceph-mon[48597]: pgmap v862: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:33.171 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:49:33.172 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:49:33.199 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:49:33.200 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:49:33.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:33 vm08.local ceph-mon[50506]: pgmap v863: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:49:33.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:33 vm01.local ceph-mon[48597]: pgmap v863: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:49:35.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:35 vm08.local ceph-mon[50506]: pgmap v864: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:35.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:35 vm01.local ceph-mon[48597]: pgmap v864: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:37.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:37 vm08.local ceph-mon[50506]: pgmap v865: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:49:37.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:37 vm01.local ceph-mon[48597]: pgmap v865: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:49:38.201 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:49:38.202 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:49:38.228 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:49:38.229 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:49:39.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:39 vm08.local ceph-mon[50506]: pgmap v866: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:39.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:39 vm01.local ceph-mon[48597]: pgmap v866: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:41.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:41 vm08.local ceph-mon[50506]: pgmap v867: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:41.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:41 vm01.local ceph-mon[48597]: pgmap v867: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:43.230 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:49:43.231 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:49:43.257 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:49:43.258 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:49:43.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:43 vm08.local ceph-mon[50506]: pgmap v868: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:49:43.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:43 vm01.local ceph-mon[48597]: pgmap v868: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:49:45.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:45 vm08.local ceph-mon[50506]: pgmap v869: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:45.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:45 vm01.local ceph-mon[48597]: pgmap v869: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:47.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:47 vm08.local ceph-mon[50506]: pgmap v870: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:49:47.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:47 vm01.local ceph-mon[48597]: pgmap v870: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:49:48.259 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:49:48.260 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:49:48.288 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:49:48.289 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:49:49.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:49 vm08.local ceph-mon[50506]: pgmap v871: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:49.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:49 vm01.local ceph-mon[48597]: pgmap v871: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:51.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:51 vm08.local ceph-mon[50506]: pgmap v872: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:51.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:51 vm01.local ceph-mon[48597]: pgmap v872: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:53.290 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:49:53.291 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:49:53.318 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:49:53.318 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:49:53.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:53 vm08.local ceph-mon[50506]: pgmap v873: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:49:53.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:53 vm01.local ceph-mon[48597]: pgmap v873: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:49:55.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:55 vm08.local ceph-mon[50506]: pgmap v874: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:55.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:55 vm01.local ceph-mon[48597]: pgmap v874: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:57.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:57 vm08.local ceph-mon[50506]: pgmap v875: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:49:57.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:57 vm01.local ceph-mon[48597]: pgmap v875: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:49:58.320 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:49:58.320 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:49:58.347 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:49:58.347 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:49:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:49:59 vm08.local ceph-mon[50506]: pgmap v876: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:49:59.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:49:59 vm01.local ceph-mon[48597]: pgmap v876: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:00.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:00 vm08.local ceph-mon[50506]: overall HEALTH_OK 2026-03-10T10:50:00.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:00 vm01.local ceph-mon[48597]: overall HEALTH_OK 2026-03-10T10:50:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:01 vm08.local ceph-mon[50506]: pgmap v877: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:01.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:01 vm01.local ceph-mon[48597]: pgmap v877: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:03.348 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:50:03.349 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:50:03.378 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:50:03.378 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:50:03.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:03 vm08.local ceph-mon[50506]: pgmap v878: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:50:03.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:03 vm01.local ceph-mon[48597]: pgmap v878: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:50:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:05 vm08.local ceph-mon[50506]: pgmap v879: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:05 vm01.local ceph-mon[48597]: pgmap v879: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:07.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:07 vm08.local ceph-mon[50506]: pgmap v880: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:50:07.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:07 vm01.local ceph-mon[48597]: pgmap v880: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:50:08.379 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:50:08.380 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:50:08.406 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:50:08.407 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:50:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:09 vm08.local ceph-mon[50506]: pgmap v881: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:09 vm01.local ceph-mon[48597]: pgmap v881: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:50:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:50:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:50:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:50:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:11 vm08.local ceph-mon[50506]: pgmap v882: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:11 vm01.local ceph-mon[48597]: pgmap v882: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:13.408 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:50:13.408 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:50:13.437 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:50:13.437 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:50:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:13 vm08.local ceph-mon[50506]: pgmap v883: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:50:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:13 vm01.local ceph-mon[48597]: pgmap v883: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:50:14.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:14 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:50:14.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:14 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:50:14.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:14 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:50:14.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:14 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:50:14.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:14 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:50:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:14 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:50:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:14 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:50:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:14 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:50:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:14 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:50:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:14 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:50:15.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:15 vm08.local ceph-mon[50506]: pgmap v884: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:15.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:15 vm01.local ceph-mon[48597]: pgmap v884: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:17.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:17 vm08.local ceph-mon[50506]: pgmap v885: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:50:17.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:17 vm01.local ceph-mon[48597]: pgmap v885: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:50:18.439 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:50:18.439 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:50:18.486 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:50:18.486 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:50:19.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:19 vm08.local ceph-mon[50506]: pgmap v886: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:19.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:19 vm01.local ceph-mon[48597]: pgmap v886: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:21.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:21 vm08.local ceph-mon[50506]: pgmap v887: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:21.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:21 vm01.local ceph-mon[48597]: pgmap v887: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:23.488 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:50:23.488 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:50:23.516 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:50:23.516 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:50:23.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:23 vm08.local ceph-mon[50506]: pgmap v888: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:50:23.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:23 vm01.local ceph-mon[48597]: pgmap v888: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:50:25.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:25 vm08.local ceph-mon[50506]: pgmap v889: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:25 vm01.local ceph-mon[48597]: pgmap v889: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:27.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:27 vm08.local ceph-mon[50506]: pgmap v890: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:50:27.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:27 vm01.local ceph-mon[48597]: pgmap v890: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:50:28.518 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:50:28.518 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:50:28.544 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:50:28.544 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:50:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:29 vm08.local ceph-mon[50506]: pgmap v891: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:29 vm01.local ceph-mon[48597]: pgmap v891: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:31.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:31 vm08.local ceph-mon[50506]: pgmap v892: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:31.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:31 vm01.local ceph-mon[48597]: pgmap v892: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:33.546 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:50:33.546 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:50:33.571 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:50:33.572 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:50:33.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:33 vm08.local ceph-mon[50506]: pgmap v893: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:50:33.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:33 vm01.local ceph-mon[48597]: pgmap v893: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:50:35.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:35 vm08.local ceph-mon[50506]: pgmap v894: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:35.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:35 vm01.local ceph-mon[48597]: pgmap v894: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:37.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:37 vm08.local ceph-mon[50506]: pgmap v895: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:50:37.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:37 vm01.local ceph-mon[48597]: pgmap v895: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:50:38.573 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:50:38.574 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:50:38.599 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:50:38.599 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:50:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:39 vm08.local ceph-mon[50506]: pgmap v896: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:39.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:39 vm01.local ceph-mon[48597]: pgmap v896: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:41.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:41 vm08.local ceph-mon[50506]: pgmap v897: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:41.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:41 vm01.local ceph-mon[48597]: pgmap v897: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:43.601 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:50:43.601 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:50:43.629 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:50:43.629 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:50:43.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:43 vm08.local ceph-mon[50506]: pgmap v898: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:50:43.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:43 vm01.local ceph-mon[48597]: pgmap v898: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:50:45.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:45 vm08.local ceph-mon[50506]: pgmap v899: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:45.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:45 vm01.local ceph-mon[48597]: pgmap v899: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:47.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:47 vm08.local ceph-mon[50506]: pgmap v900: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:50:47.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:47 vm01.local ceph-mon[48597]: pgmap v900: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:50:48.631 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:50:48.631 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:50:48.656 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:50:48.657 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:50:49.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:49 vm08.local ceph-mon[50506]: pgmap v901: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:49.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:49 vm01.local ceph-mon[48597]: pgmap v901: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:51.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:51 vm08.local ceph-mon[50506]: pgmap v902: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:51.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:51 vm01.local ceph-mon[48597]: pgmap v902: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:53.658 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:50:53.659 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:50:53.685 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:50:53.686 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:50:53.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:53 vm08.local ceph-mon[50506]: pgmap v903: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:50:53.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:53 vm01.local ceph-mon[48597]: pgmap v903: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:50:55.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:55 vm08.local ceph-mon[50506]: pgmap v904: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:55.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:55 vm01.local ceph-mon[48597]: pgmap v904: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:57.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:57 vm08.local ceph-mon[50506]: pgmap v905: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:50:57.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:57 vm01.local ceph-mon[48597]: pgmap v905: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:50:58.687 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:50:58.688 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:50:58.713 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:50:58.714 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:50:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:50:59 vm08.local ceph-mon[50506]: pgmap v906: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:50:59.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:50:59 vm01.local ceph-mon[48597]: pgmap v906: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:01.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:01 vm08.local ceph-mon[50506]: pgmap v907: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:01.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:01 vm01.local ceph-mon[48597]: pgmap v907: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:03.715 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:51:03.716 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:51:03.743 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:51:03.743 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:51:03.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:03 vm08.local ceph-mon[50506]: pgmap v908: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:51:03.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:03 vm01.local ceph-mon[48597]: pgmap v908: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:51:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:05 vm08.local ceph-mon[50506]: pgmap v909: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:05 vm01.local ceph-mon[48597]: pgmap v909: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:07.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:07 vm08.local ceph-mon[50506]: pgmap v910: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:51:07.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:07 vm01.local ceph-mon[48597]: pgmap v910: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:51:08.745 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:51:08.746 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:51:08.773 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:51:08.774 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:51:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:09 vm08.local ceph-mon[50506]: pgmap v911: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:09 vm01.local ceph-mon[48597]: pgmap v911: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:51:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:51:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:51:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:51:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:11 vm08.local ceph-mon[50506]: pgmap v912: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:11 vm01.local ceph-mon[48597]: pgmap v912: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:13.775 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:51:13.776 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:51:13.802 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:51:13.803 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:51:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:13 vm08.local ceph-mon[50506]: pgmap v913: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:51:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:13 vm01.local ceph-mon[48597]: pgmap v913: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:51:14.476 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:14 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:51:14.476 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:14 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:51:14.476 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:14 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:51:14.476 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:14 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:51:14.476 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:14 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:51:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:14 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:51:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:14 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:51:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:14 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:51:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:14 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:51:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:14 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:51:15.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:15 vm08.local ceph-mon[50506]: pgmap v914: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:15.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:15 vm01.local ceph-mon[48597]: pgmap v914: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:17.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:17 vm08.local ceph-mon[50506]: pgmap v915: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:51:17.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:17 vm01.local ceph-mon[48597]: pgmap v915: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:51:18.804 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:51:18.804 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:51:18.831 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:51:18.831 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:51:19.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:19 vm08.local ceph-mon[50506]: pgmap v916: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:19.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:19 vm01.local ceph-mon[48597]: pgmap v916: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:21.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:21 vm08.local ceph-mon[50506]: pgmap v917: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:21.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:21 vm01.local ceph-mon[48597]: pgmap v917: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:23.832 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:51:23.833 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:51:23.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:23 vm08.local ceph-mon[50506]: pgmap v918: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:51:23.860 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:51:23.860 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:51:23.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:23 vm01.local ceph-mon[48597]: pgmap v918: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:51:25.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:25 vm08.local ceph-mon[50506]: pgmap v919: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:25 vm01.local ceph-mon[48597]: pgmap v919: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:27.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:27 vm08.local ceph-mon[50506]: pgmap v920: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:51:27.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:27 vm01.local ceph-mon[48597]: pgmap v920: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:51:28.861 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:51:28.862 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:51:28.888 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:51:28.888 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:51:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:29 vm08.local ceph-mon[50506]: pgmap v921: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:29 vm01.local ceph-mon[48597]: pgmap v921: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:31.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:31 vm08.local ceph-mon[50506]: pgmap v922: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:31.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:31 vm01.local ceph-mon[48597]: pgmap v922: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:33.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:33 vm08.local ceph-mon[50506]: pgmap v923: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:51:33.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:33 vm01.local ceph-mon[48597]: pgmap v923: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:51:33.889 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:51:33.890 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:51:33.915 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:51:33.916 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:51:35.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:35 vm08.local ceph-mon[50506]: pgmap v924: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:35.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:35 vm01.local ceph-mon[48597]: pgmap v924: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:37.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:37 vm08.local ceph-mon[50506]: pgmap v925: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:51:37.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:37 vm01.local ceph-mon[48597]: pgmap v925: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:51:38.917 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:51:38.918 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:51:38.943 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:51:38.944 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:51:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:39 vm08.local ceph-mon[50506]: pgmap v926: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:39.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:39 vm01.local ceph-mon[48597]: pgmap v926: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:41.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:41 vm08.local ceph-mon[50506]: pgmap v927: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:41.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:41 vm01.local ceph-mon[48597]: pgmap v927: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:43.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:43 vm08.local ceph-mon[50506]: pgmap v928: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:51:43.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:43 vm01.local ceph-mon[48597]: pgmap v928: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:51:43.945 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:51:43.946 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:51:43.971 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:51:43.972 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:51:45.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:45 vm08.local ceph-mon[50506]: pgmap v929: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:45.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:45 vm01.local ceph-mon[48597]: pgmap v929: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:47.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:47 vm08.local ceph-mon[50506]: pgmap v930: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:51:47.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:47 vm01.local ceph-mon[48597]: pgmap v930: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:51:48.973 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:51:48.974 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:51:49.000 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:51:49.000 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:51:49.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:49 vm08.local ceph-mon[50506]: pgmap v931: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:49.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:49 vm01.local ceph-mon[48597]: pgmap v931: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:51.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:51 vm08.local ceph-mon[50506]: pgmap v932: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:51.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:51 vm01.local ceph-mon[48597]: pgmap v932: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:53.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:53 vm08.local ceph-mon[50506]: pgmap v933: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:51:53.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:53 vm01.local ceph-mon[48597]: pgmap v933: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:51:54.001 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:51:54.002 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:51:54.028 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:51:54.029 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:51:55.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:55 vm08.local ceph-mon[50506]: pgmap v934: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:55.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:55 vm01.local ceph-mon[48597]: pgmap v934: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:57.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:57 vm08.local ceph-mon[50506]: pgmap v935: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:51:57.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:57 vm01.local ceph-mon[48597]: pgmap v935: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:51:59.030 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:51:59.030 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:51:59.055 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:51:59.056 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:51:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:51:59 vm08.local ceph-mon[50506]: pgmap v936: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:51:59.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:51:59 vm01.local ceph-mon[48597]: pgmap v936: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:01.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:01 vm08.local ceph-mon[50506]: pgmap v937: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:01.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:01 vm01.local ceph-mon[48597]: pgmap v937: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:03.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:03 vm08.local ceph-mon[50506]: pgmap v938: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:52:03.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:03 vm01.local ceph-mon[48597]: pgmap v938: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:52:04.057 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:52:04.057 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:52:04.092 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:52:04.092 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:52:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:05 vm08.local ceph-mon[50506]: pgmap v939: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:05 vm01.local ceph-mon[48597]: pgmap v939: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:07.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:07 vm08.local ceph-mon[50506]: pgmap v940: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:52:07.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:07 vm01.local ceph-mon[48597]: pgmap v940: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:52:09.094 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:52:09.094 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:52:09.120 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:52:09.120 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:52:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:09 vm08.local ceph-mon[50506]: pgmap v941: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:09 vm01.local ceph-mon[48597]: pgmap v941: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:52:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:52:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:52:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:52:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:11 vm08.local ceph-mon[50506]: pgmap v942: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:11 vm01.local ceph-mon[48597]: pgmap v942: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:13 vm08.local ceph-mon[50506]: pgmap v943: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:52:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:13 vm01.local ceph-mon[48597]: pgmap v943: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:52:14.122 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:52:14.122 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:52:14.149 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:52:14.150 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:52:14.803 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:14 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:52:14.803 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:14 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:52:14.803 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:14 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:52:14.844 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:14 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:52:14.844 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:14 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:52:14.844 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:14 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:52:16.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:15 vm08.local ceph-mon[50506]: pgmap v944: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:16.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:15 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:52:16.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:15 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:52:16.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:15 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:52:16.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:15 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:52:16.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:15 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:52:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:15 vm01.local ceph-mon[48597]: pgmap v944: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:15 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:52:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:15 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:52:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:15 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:52:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:15 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:52:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:15 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:52:18.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:17 vm08.local ceph-mon[50506]: pgmap v945: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:52:18.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:17 vm01.local ceph-mon[48597]: pgmap v945: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:52:19.151 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:52:19.151 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:52:19.177 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:52:19.177 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:52:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:19 vm08.local ceph-mon[50506]: pgmap v946: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:19 vm01.local ceph-mon[48597]: pgmap v946: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:21.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:21 vm08.local ceph-mon[50506]: pgmap v947: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:21.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:21 vm01.local ceph-mon[48597]: pgmap v947: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:23.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:23 vm08.local ceph-mon[50506]: pgmap v948: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:52:23.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:23 vm01.local ceph-mon[48597]: pgmap v948: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:52:24.179 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:52:24.179 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:52:24.204 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:52:24.205 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:52:25.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:25 vm08.local ceph-mon[50506]: pgmap v949: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:25 vm01.local ceph-mon[48597]: pgmap v949: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:27.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:27 vm08.local ceph-mon[50506]: pgmap v950: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:52:27.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:27 vm01.local ceph-mon[48597]: pgmap v950: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:52:29.206 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:52:29.207 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:52:29.232 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:52:29.233 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:52:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:29 vm08.local ceph-mon[50506]: pgmap v951: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:29 vm01.local ceph-mon[48597]: pgmap v951: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:31.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:31 vm08.local ceph-mon[50506]: pgmap v952: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:31.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:31 vm01.local ceph-mon[48597]: pgmap v952: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:33.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:33 vm08.local ceph-mon[50506]: pgmap v953: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:52:33.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:33 vm01.local ceph-mon[48597]: pgmap v953: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:52:34.234 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:52:34.234 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:52:34.260 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:52:34.261 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:52:35.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:35 vm08.local ceph-mon[50506]: pgmap v954: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:35.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:35 vm01.local ceph-mon[48597]: pgmap v954: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:37.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:37 vm08.local ceph-mon[50506]: pgmap v955: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:52:37.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:37 vm01.local ceph-mon[48597]: pgmap v955: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:52:39.262 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:52:39.263 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:52:39.289 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:52:39.289 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:52:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:39 vm08.local ceph-mon[50506]: pgmap v956: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:39.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:39 vm01.local ceph-mon[48597]: pgmap v956: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:41.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:41 vm08.local ceph-mon[50506]: pgmap v957: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:41.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:41 vm01.local ceph-mon[48597]: pgmap v957: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:43.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:43 vm08.local ceph-mon[50506]: pgmap v958: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:52:43.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:43 vm01.local ceph-mon[48597]: pgmap v958: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:52:44.291 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:52:44.291 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:52:44.318 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:52:44.319 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:52:45.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:45 vm08.local ceph-mon[50506]: pgmap v959: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:45.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:45 vm01.local ceph-mon[48597]: pgmap v959: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:47.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:47 vm08.local ceph-mon[50506]: pgmap v960: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:52:47.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:47 vm01.local ceph-mon[48597]: pgmap v960: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:52:49.320 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:52:49.321 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:52:49.347 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:52:49.347 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:52:49.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:49 vm08.local ceph-mon[50506]: pgmap v961: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:49.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:49 vm01.local ceph-mon[48597]: pgmap v961: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:51.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:51 vm01.local ceph-mon[48597]: pgmap v962: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:51.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:51 vm08.local ceph-mon[50506]: pgmap v962: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:53.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:53 vm08.local ceph-mon[50506]: pgmap v963: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:52:53.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:53 vm01.local ceph-mon[48597]: pgmap v963: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:52:54.348 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:52:54.349 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:52:54.374 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:52:54.374 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:52:55.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:55 vm08.local ceph-mon[50506]: pgmap v964: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:55.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:55 vm01.local ceph-mon[48597]: pgmap v964: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:57.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:57 vm08.local ceph-mon[50506]: pgmap v965: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:52:57.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:57 vm01.local ceph-mon[48597]: pgmap v965: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:52:59.376 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:52:59.376 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:52:59.401 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:52:59.402 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:52:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:52:59 vm08.local ceph-mon[50506]: pgmap v966: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:52:59.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:52:59 vm01.local ceph-mon[48597]: pgmap v966: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:01.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:01 vm01.local ceph-mon[48597]: pgmap v967: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:01 vm08.local ceph-mon[50506]: pgmap v967: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:03.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:03 vm08.local ceph-mon[50506]: pgmap v968: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:53:03.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:03 vm01.local ceph-mon[48597]: pgmap v968: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:53:04.403 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:53:04.404 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:53:04.435 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:53:04.436 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:53:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:05 vm08.local ceph-mon[50506]: pgmap v969: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:05 vm01.local ceph-mon[48597]: pgmap v969: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:07.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:07 vm08.local ceph-mon[50506]: pgmap v970: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:53:07.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:07 vm01.local ceph-mon[48597]: pgmap v970: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:53:09.437 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:53:09.438 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:53:09.463 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:53:09.464 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:53:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:09 vm08.local ceph-mon[50506]: pgmap v971: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:09 vm01.local ceph-mon[48597]: pgmap v971: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:53:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:53:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:53:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:53:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:11 vm08.local ceph-mon[50506]: pgmap v972: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:11 vm01.local ceph-mon[48597]: pgmap v972: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:13 vm08.local ceph-mon[50506]: pgmap v973: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:53:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:13 vm01.local ceph-mon[48597]: pgmap v973: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:53:14.465 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:53:14.465 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:53:14.490 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:53:14.490 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:53:15.605 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:15 vm01.local ceph-mon[48597]: pgmap v974: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:15.605 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:15 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:53:15.605 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:15 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:53:15.605 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:15 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:53:15.606 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:15 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:53:15.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:15 vm08.local ceph-mon[50506]: pgmap v974: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:15.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:15 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:53:15.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:15 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:53:15.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:15 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:53:15.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:15 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:53:17.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:17 vm08.local ceph-mon[50506]: pgmap v975: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:53:17.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:17 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:53:17.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:17 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:53:17.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:17 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:53:17.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:17 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:53:17.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:17 vm01.local ceph-mon[48597]: pgmap v975: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:53:17.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:17 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:53:17.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:17 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:53:17.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:17 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:53:17.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:17 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:53:19.492 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:53:19.493 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:53:19.521 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:53:19.521 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:53:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:19 vm08.local ceph-mon[50506]: pgmap v976: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:19 vm01.local ceph-mon[48597]: pgmap v976: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:21.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:21 vm08.local ceph-mon[50506]: pgmap v977: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:21 vm01.local ceph-mon[48597]: pgmap v977: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:23.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:23 vm08.local ceph-mon[50506]: pgmap v978: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:53:23.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:23 vm01.local ceph-mon[48597]: pgmap v978: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:53:24.522 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:53:24.523 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:53:24.549 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:53:24.550 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:53:25.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:25 vm08.local ceph-mon[50506]: pgmap v979: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:25 vm01.local ceph-mon[48597]: pgmap v979: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:27.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:27 vm08.local ceph-mon[50506]: pgmap v980: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:53:27.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:27 vm01.local ceph-mon[48597]: pgmap v980: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:53:29.551 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:53:29.552 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:53:29.578 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:53:29.578 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:53:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:29 vm08.local ceph-mon[50506]: pgmap v981: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:29 vm01.local ceph-mon[48597]: pgmap v981: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:31.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:31 vm08.local ceph-mon[50506]: pgmap v982: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:31.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:31 vm01.local ceph-mon[48597]: pgmap v982: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:33.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:33 vm08.local ceph-mon[50506]: pgmap v983: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:53:33.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:33 vm01.local ceph-mon[48597]: pgmap v983: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:53:34.580 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:53:34.581 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:53:34.607 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:53:34.607 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:53:35.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:35 vm08.local ceph-mon[50506]: pgmap v984: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:35.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:35 vm01.local ceph-mon[48597]: pgmap v984: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:37.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:37 vm08.local ceph-mon[50506]: pgmap v985: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:53:37.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:37 vm01.local ceph-mon[48597]: pgmap v985: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:53:39.609 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:53:39.610 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:53:39.637 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:53:39.638 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:53:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:39 vm08.local ceph-mon[50506]: pgmap v986: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:39.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:39 vm01.local ceph-mon[48597]: pgmap v986: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:41.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:41 vm08.local ceph-mon[50506]: pgmap v987: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:41.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:41 vm01.local ceph-mon[48597]: pgmap v987: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:43.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:43 vm08.local ceph-mon[50506]: pgmap v988: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:53:43.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:43 vm01.local ceph-mon[48597]: pgmap v988: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:53:44.639 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:53:44.640 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:53:44.667 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:53:44.668 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:53:45.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:45 vm08.local ceph-mon[50506]: pgmap v989: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:45.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:45 vm01.local ceph-mon[48597]: pgmap v989: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:47.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:47 vm08.local ceph-mon[50506]: pgmap v990: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:53:47.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:47 vm01.local ceph-mon[48597]: pgmap v990: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:53:49.669 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:53:49.669 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:53:49.695 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:53:49.696 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:53:49.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:49 vm08.local ceph-mon[50506]: pgmap v991: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:49.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:49 vm01.local ceph-mon[48597]: pgmap v991: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:51.743 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:51 vm01.local ceph-mon[48597]: pgmap v992: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:51.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:51 vm08.local ceph-mon[50506]: pgmap v992: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:53.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:53 vm08.local ceph-mon[50506]: pgmap v993: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:53:53.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:53 vm01.local ceph-mon[48597]: pgmap v993: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:53:54.697 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:53:54.698 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:53:54.725 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:53:54.725 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:53:55.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:55 vm08.local ceph-mon[50506]: pgmap v994: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:55.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:55 vm01.local ceph-mon[48597]: pgmap v994: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:57.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:57 vm08.local ceph-mon[50506]: pgmap v995: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:53:57.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:57 vm01.local ceph-mon[48597]: pgmap v995: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:53:59.727 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:53:59.727 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:53:59.752 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:53:59.753 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:53:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:53:59 vm08.local ceph-mon[50506]: pgmap v996: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:53:59.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:53:59 vm01.local ceph-mon[48597]: pgmap v996: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:01 vm08.local ceph-mon[50506]: pgmap v997: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:01.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:01 vm01.local ceph-mon[48597]: pgmap v997: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:03.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:03 vm08.local ceph-mon[50506]: pgmap v998: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:54:03.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:03 vm01.local ceph-mon[48597]: pgmap v998: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:54:04.754 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:54:04.755 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:54:04.780 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:54:04.780 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:54:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:05 vm08.local ceph-mon[50506]: pgmap v999: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:05 vm01.local ceph-mon[48597]: pgmap v999: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:07.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:07 vm08.local ceph-mon[50506]: pgmap v1000: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:54:07.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:07 vm01.local ceph-mon[48597]: pgmap v1000: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:54:09.782 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:54:09.782 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:54:09.807 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:54:09.808 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:54:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:09 vm08.local ceph-mon[50506]: pgmap v1001: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:09 vm01.local ceph-mon[48597]: pgmap v1001: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:54:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:54:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:54:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:54:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:11 vm08.local ceph-mon[50506]: pgmap v1002: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:11 vm01.local ceph-mon[48597]: pgmap v1002: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:13.647 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:13 vm08.local ceph-mon[50506]: pgmap v1003: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:54:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:13 vm01.local ceph-mon[48597]: pgmap v1003: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:54:14.809 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:54:14.810 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:54:14.836 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:54:14.836 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:54:15.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:15 vm08.local ceph-mon[50506]: pgmap v1004: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:15.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:15 vm01.local ceph-mon[48597]: pgmap v1004: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:17 vm08.local ceph-mon[50506]: pgmap v1005: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:54:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:17 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:54:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:17 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:54:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:17 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:54:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:17 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:54:17.790 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:17 vm01.local ceph-mon[48597]: pgmap v1005: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:54:17.790 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:17 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:54:17.790 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:17 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:54:17.790 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:17 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:54:17.790 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:17 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:54:19.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:18 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:54:19.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:18 vm08.local ceph-mon[50506]: pgmap v1006: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:19.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:18 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:54:19.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:18 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:54:19.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:18 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:54:19.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:18 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:54:19.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:18 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:54:19.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:18 vm01.local ceph-mon[48597]: pgmap v1006: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:19.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:18 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:54:19.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:18 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:54:19.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:18 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:54:19.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:18 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:54:19.838 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:54:19.838 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:54:19.908 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:54:19.909 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:54:21.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:21 vm08.local ceph-mon[50506]: pgmap v1007: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:21 vm01.local ceph-mon[48597]: pgmap v1007: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:23.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:23 vm08.local ceph-mon[50506]: pgmap v1008: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:54:23.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:23 vm01.local ceph-mon[48597]: pgmap v1008: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:54:24.911 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:54:24.911 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:54:24.939 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:54:24.940 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:54:25.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:25 vm08.local ceph-mon[50506]: pgmap v1009: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:25 vm01.local ceph-mon[48597]: pgmap v1009: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:27.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:27 vm08.local ceph-mon[50506]: pgmap v1010: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:54:27.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:27 vm01.local ceph-mon[48597]: pgmap v1010: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:54:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:29 vm08.local ceph-mon[50506]: pgmap v1011: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:29 vm01.local ceph-mon[48597]: pgmap v1011: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:29.941 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:54:29.942 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:54:29.968 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:54:29.968 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:54:31.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:31 vm08.local ceph-mon[50506]: pgmap v1012: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:31.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:31 vm01.local ceph-mon[48597]: pgmap v1012: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:33.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:33 vm08.local ceph-mon[50506]: pgmap v1013: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:54:33.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:33 vm01.local ceph-mon[48597]: pgmap v1013: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:54:34.969 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:54:34.970 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:54:34.995 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:54:34.996 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:54:35.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:35 vm08.local ceph-mon[50506]: pgmap v1014: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:35.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:35 vm01.local ceph-mon[48597]: pgmap v1014: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:37.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:37 vm01.local ceph-mon[48597]: pgmap v1015: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:54:38.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:37 vm08.local ceph-mon[50506]: pgmap v1015: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:54:39.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:39 vm01.local ceph-mon[48597]: pgmap v1016: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:39.997 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:54:39.997 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:54:40.023 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:54:40.024 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:54:40.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:39 vm08.local ceph-mon[50506]: pgmap v1016: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:41.646 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:41 vm01.local ceph-mon[48597]: pgmap v1017: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:42.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:41 vm08.local ceph-mon[50506]: pgmap v1017: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:43.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:43 vm01.local ceph-mon[48597]: pgmap v1018: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:54:43.896 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:43 vm08.local ceph-mon[50506]: pgmap v1018: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:54:45.025 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:54:45.025 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:54:45.053 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:54:45.053 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:54:46.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:45 vm08.local ceph-mon[50506]: pgmap v1019: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:46.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:45 vm01.local ceph-mon[48597]: pgmap v1019: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:48.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:47 vm08.local ceph-mon[50506]: pgmap v1020: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:54:48.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:47 vm01.local ceph-mon[48597]: pgmap v1020: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:54:50.055 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:54:50.055 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:54:50.082 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:54:50.082 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:54:50.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:49 vm08.local ceph-mon[50506]: pgmap v1021: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:50.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:49 vm01.local ceph-mon[48597]: pgmap v1021: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:52.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:51 vm08.local ceph-mon[50506]: pgmap v1022: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:52.116 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:51 vm01.local ceph-mon[48597]: pgmap v1022: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:54.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:53 vm08.local ceph-mon[50506]: pgmap v1023: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:54:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:53 vm01.local ceph-mon[48597]: pgmap v1023: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:54:55.084 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:54:55.084 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:54:55.114 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:54:55.115 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:54:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:55 vm08.local ceph-mon[50506]: pgmap v1024: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:56.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:55 vm01.local ceph-mon[48597]: pgmap v1024: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:54:58.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:57 vm08.local ceph-mon[50506]: pgmap v1025: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:54:58.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:57 vm01.local ceph-mon[48597]: pgmap v1025: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:55:00.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:54:59 vm08.local ceph-mon[50506]: pgmap v1026: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:00.116 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:55:00.116 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:55:00.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:54:59 vm01.local ceph-mon[48597]: pgmap v1026: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:00.144 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:55:00.144 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:55:02.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:01 vm08.local ceph-mon[50506]: pgmap v1027: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:02.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:01 vm01.local ceph-mon[48597]: pgmap v1027: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:04.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:03 vm08.local ceph-mon[50506]: pgmap v1028: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:55:04.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:03 vm01.local ceph-mon[48597]: pgmap v1028: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:55:05.146 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:55:05.146 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:55:05.172 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:55:05.172 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:55:06.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:05 vm08.local ceph-mon[50506]: pgmap v1029: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:06.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:05 vm01.local ceph-mon[48597]: pgmap v1029: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:08.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:07 vm08.local ceph-mon[50506]: pgmap v1030: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:55:08.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:07 vm01.local ceph-mon[48597]: pgmap v1030: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:55:10.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:09 vm08.local ceph-mon[50506]: pgmap v1031: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:10.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:09 vm01.local ceph-mon[48597]: pgmap v1031: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:10.173 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:55:10.174 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:55:10.200 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:55:10.201 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:55:10.992 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:55:10.992 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:55:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:55:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:55:12.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:11 vm08.local ceph-mon[50506]: pgmap v1032: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:12.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:11 vm01.local ceph-mon[48597]: pgmap v1032: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:14.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:13 vm08.local ceph-mon[50506]: pgmap v1033: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:55:14.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:13 vm01.local ceph-mon[48597]: pgmap v1033: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:55:15.203 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:55:15.204 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:55:15.231 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:55:15.232 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:55:16.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:15 vm08.local ceph-mon[50506]: pgmap v1034: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:15 vm01.local ceph-mon[48597]: pgmap v1034: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:18.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:17 vm08.local ceph-mon[50506]: pgmap v1035: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:55:18.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:17 vm01.local ceph-mon[48597]: pgmap v1035: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:55:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:19 vm08.local ceph-mon[50506]: pgmap v1036: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:19 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:55:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:19 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:55:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:19 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:55:20.095 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:19 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:55:20.095 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:19 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:55:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:19 vm01.local ceph-mon[48597]: pgmap v1036: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:19 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:55:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:19 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:55:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:19 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:55:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:19 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:55:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:19 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:55:20.233 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:55:20.234 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:55:20.259 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:55:20.260 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:55:22.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:21 vm08.local ceph-mon[50506]: pgmap v1037: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:21 vm01.local ceph-mon[48597]: pgmap v1037: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:22.932 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:22 vm01.local ceph-mon[48597]: pgmap v1038: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:55:23.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:22 vm08.local ceph-mon[50506]: pgmap v1038: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:55:25.261 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:55:25.262 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:55:25.291 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:55:25.291 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:55:25.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:25 vm08.local ceph-mon[50506]: pgmap v1039: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:25 vm01.local ceph-mon[48597]: pgmap v1039: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:27.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:27 vm08.local ceph-mon[50506]: pgmap v1040: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:55:27.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:27 vm01.local ceph-mon[48597]: pgmap v1040: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:55:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:29 vm08.local ceph-mon[50506]: pgmap v1041: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:29 vm01.local ceph-mon[48597]: pgmap v1041: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:30.292 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:55:30.293 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:55:30.319 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:55:30.320 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:55:31.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:31 vm08.local ceph-mon[50506]: pgmap v1042: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:31.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:31 vm01.local ceph-mon[48597]: pgmap v1042: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:33.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:33 vm08.local ceph-mon[50506]: pgmap v1043: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:55:33.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:33 vm01.local ceph-mon[48597]: pgmap v1043: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:55:35.322 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:55:35.322 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:55:35.348 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:55:35.349 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:55:35.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:35 vm08.local ceph-mon[50506]: pgmap v1044: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:35.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:35 vm01.local ceph-mon[48597]: pgmap v1044: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:37.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:37 vm08.local ceph-mon[50506]: pgmap v1045: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:55:37.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:37 vm01.local ceph-mon[48597]: pgmap v1045: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:55:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:39 vm08.local ceph-mon[50506]: pgmap v1046: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:39.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:39 vm01.local ceph-mon[48597]: pgmap v1046: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:40.351 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:55:40.351 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:55:40.377 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:55:40.377 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:55:41.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:41 vm08.local ceph-mon[50506]: pgmap v1047: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:41.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:41 vm01.local ceph-mon[48597]: pgmap v1047: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:43.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:43 vm08.local ceph-mon[50506]: pgmap v1048: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:55:43.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:43 vm01.local ceph-mon[48597]: pgmap v1048: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:55:45.379 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:55:45.379 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:55:45.405 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:55:45.406 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:55:45.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:45 vm08.local ceph-mon[50506]: pgmap v1049: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:45.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:45 vm01.local ceph-mon[48597]: pgmap v1049: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:47.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:47 vm08.local ceph-mon[50506]: pgmap v1050: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:55:47.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:47 vm01.local ceph-mon[48597]: pgmap v1050: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:55:49.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:49 vm01.local ceph-mon[48597]: pgmap v1051: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:50.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:49 vm08.local ceph-mon[50506]: pgmap v1051: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:50.407 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:55:50.408 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:55:50.434 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:55:50.434 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:55:51.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:51 vm01.local ceph-mon[48597]: pgmap v1052: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:52.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:51 vm08.local ceph-mon[50506]: pgmap v1052: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:53.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:53 vm01.local ceph-mon[48597]: pgmap v1053: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:55:54.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:53 vm08.local ceph-mon[50506]: pgmap v1053: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:55:55.436 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:55:55.436 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:55:55.467 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:55:55.468 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:55:55.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:55 vm01.local ceph-mon[48597]: pgmap v1054: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:55 vm08.local ceph-mon[50506]: pgmap v1054: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:55:58.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:57 vm08.local ceph-mon[50506]: pgmap v1055: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:55:58.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:57 vm01.local ceph-mon[48597]: pgmap v1055: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:56:00.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:55:59 vm08.local ceph-mon[50506]: pgmap v1056: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:00.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:55:59 vm01.local ceph-mon[48597]: pgmap v1056: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:00.469 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:56:00.470 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:56:00.495 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:56:00.496 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:56:02.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:01 vm08.local ceph-mon[50506]: pgmap v1057: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:02.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:01 vm01.local ceph-mon[48597]: pgmap v1057: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:04.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:03 vm08.local ceph-mon[50506]: pgmap v1058: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:56:04.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:03 vm01.local ceph-mon[48597]: pgmap v1058: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:56:05.497 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:56:05.498 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:56:05.526 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:56:05.526 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:56:06.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:05 vm08.local ceph-mon[50506]: pgmap v1059: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:06.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:05 vm01.local ceph-mon[48597]: pgmap v1059: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:08.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:07 vm08.local ceph-mon[50506]: pgmap v1060: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:08.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:07 vm01.local ceph-mon[48597]: pgmap v1060: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:10.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:09 vm08.local ceph-mon[50506]: pgmap v1061: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T10:56:10.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:09 vm01.local ceph-mon[48597]: pgmap v1061: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T10:56:10.528 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:56:10.528 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:56:10.554 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:56:10.555 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:56:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:56:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:56:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:56:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:56:12.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:11 vm08.local ceph-mon[50506]: pgmap v1062: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T10:56:12.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:11 vm01.local ceph-mon[48597]: pgmap v1062: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T10:56:13.921 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:13 vm01.local ceph-mon[48597]: pgmap v1063: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:14.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:13 vm08.local ceph-mon[50506]: pgmap v1063: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:15.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:14 vm08.local ceph-mon[50506]: pgmap v1064: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T10:56:15.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:14 vm01.local ceph-mon[48597]: pgmap v1064: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T10:56:15.557 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:56:15.557 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:56:15.585 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:56:15.586 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:56:17.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:17 vm08.local ceph-mon[50506]: pgmap v1065: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:17.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:17 vm01.local ceph-mon[48597]: pgmap v1065: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:19.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:19 vm08.local ceph-mon[50506]: pgmap v1066: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:19.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:19 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:56:19.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:19 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:56:19.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:19 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:56:19.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:19 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:56:19.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:19 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:56:19.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:19 vm01.local ceph-mon[48597]: pgmap v1066: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:19.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:19 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:56:19.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:19 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:56:19.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:19 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:56:19.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:19 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:56:19.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:19 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:56:20.587 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:56:20.588 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:56:20.615 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:56:20.615 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:56:21.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:21 vm08.local ceph-mon[50506]: pgmap v1067: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:21 vm01.local ceph-mon[48597]: pgmap v1067: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:23.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:23 vm08.local ceph-mon[50506]: pgmap v1068: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:56:23.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:23 vm01.local ceph-mon[48597]: pgmap v1068: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:56:25.617 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:56:25.617 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:56:25.646 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:56:25.647 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:56:25.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:25 vm08.local ceph-mon[50506]: pgmap v1069: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:25 vm01.local ceph-mon[48597]: pgmap v1069: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:27.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:27 vm08.local ceph-mon[50506]: pgmap v1070: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:56:27.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:27 vm01.local ceph-mon[48597]: pgmap v1070: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:56:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:29 vm08.local ceph-mon[50506]: pgmap v1071: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:29 vm01.local ceph-mon[48597]: pgmap v1071: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:30.648 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:56:30.649 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:56:30.678 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:56:30.679 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:56:32.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:31 vm08.local ceph-mon[50506]: pgmap v1072: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:32.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:31 vm01.local ceph-mon[48597]: pgmap v1072: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:34.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:33 vm08.local ceph-mon[50506]: pgmap v1073: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:56:34.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:33 vm01.local ceph-mon[48597]: pgmap v1073: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:56:35.680 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:56:35.681 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:56:35.707 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:56:35.707 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:56:36.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:35 vm08.local ceph-mon[50506]: pgmap v1074: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:35 vm01.local ceph-mon[48597]: pgmap v1074: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:38.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:37 vm08.local ceph-mon[50506]: pgmap v1075: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:56:38.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:37 vm01.local ceph-mon[48597]: pgmap v1075: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:56:40.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:39 vm08.local ceph-mon[50506]: pgmap v1076: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:39 vm01.local ceph-mon[48597]: pgmap v1076: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:40.709 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:56:40.709 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:56:40.739 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:56:40.740 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:56:42.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:41 vm08.local ceph-mon[50506]: pgmap v1077: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:42.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:41 vm01.local ceph-mon[48597]: pgmap v1077: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:43.921 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:43 vm01.local ceph-mon[48597]: pgmap v1078: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:56:44.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:43 vm08.local ceph-mon[50506]: pgmap v1078: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:56:45.741 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:56:45.742 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:56:45.770 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:56:45.771 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:56:46.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:45 vm08.local ceph-mon[50506]: pgmap v1079: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:46.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:45 vm01.local ceph-mon[48597]: pgmap v1079: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:48.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:47 vm08.local ceph-mon[50506]: pgmap v1080: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:56:48.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:47 vm01.local ceph-mon[48597]: pgmap v1080: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:56:50.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:49 vm08.local ceph-mon[50506]: pgmap v1081: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:50.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:49 vm01.local ceph-mon[48597]: pgmap v1081: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:50.773 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:56:50.773 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:56:50.843 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:56:50.844 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:56:52.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:51 vm08.local ceph-mon[50506]: pgmap v1082: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:52.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:51 vm01.local ceph-mon[48597]: pgmap v1082: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:54.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:53 vm08.local ceph-mon[50506]: pgmap v1083: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:56:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:53 vm01.local ceph-mon[48597]: pgmap v1083: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:56:55.845 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:56:55.846 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:56:55.878 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:56:55.878 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:56:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:55 vm08.local ceph-mon[50506]: pgmap v1084: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:56.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:55 vm01.local ceph-mon[48597]: pgmap v1084: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:58.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:57 vm08.local ceph-mon[50506]: pgmap v1085: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:56:58.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:57 vm01.local ceph-mon[48597]: pgmap v1085: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:56:59.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:56:58 vm08.local ceph-mon[50506]: pgmap v1086: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:56:59.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:56:58 vm01.local ceph-mon[48597]: pgmap v1086: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:00.879 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:57:00.880 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:57:00.907 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:57:00.907 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:57:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:01 vm08.local ceph-mon[50506]: pgmap v1087: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:01.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:01 vm01.local ceph-mon[48597]: pgmap v1087: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:03.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:03 vm08.local ceph-mon[50506]: pgmap v1088: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:57:03.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:03 vm01.local ceph-mon[48597]: pgmap v1088: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:57:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:05 vm08.local ceph-mon[50506]: pgmap v1089: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:05 vm01.local ceph-mon[48597]: pgmap v1089: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:05.909 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:57:05.909 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:57:05.935 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:57:05.936 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:57:07.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:07 vm08.local ceph-mon[50506]: pgmap v1090: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:57:07.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:07 vm01.local ceph-mon[48597]: pgmap v1090: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:57:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:09 vm08.local ceph-mon[50506]: pgmap v1091: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:09 vm01.local ceph-mon[48597]: pgmap v1091: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:57:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:57:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:57:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:57:10.938 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:57:10.938 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:57:10.965 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:57:10.965 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:57:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:11 vm08.local ceph-mon[50506]: pgmap v1092: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:11 vm01.local ceph-mon[48597]: pgmap v1092: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:14.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:13 vm08.local ceph-mon[50506]: pgmap v1093: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:57:14.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:13 vm01.local ceph-mon[48597]: pgmap v1093: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:57:15.967 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:57:15.967 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:57:15.994 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:57:15.995 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:57:16.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:15 vm08.local ceph-mon[50506]: pgmap v1094: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:15 vm01.local ceph-mon[48597]: pgmap v1094: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:18.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:17 vm08.local ceph-mon[50506]: pgmap v1095: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:57:18.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:17 vm01.local ceph-mon[48597]: pgmap v1095: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:57:19.691 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:19 vm01.local ceph-mon[48597]: pgmap v1096: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:19.691 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:19 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:57:19.692 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:19 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:57:19.692 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:19 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:57:19.692 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:19 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:57:19.692 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:19 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:57:19.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:19 vm08.local ceph-mon[50506]: pgmap v1096: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:19.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:19 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:57:19.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:19 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:57:19.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:19 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:57:19.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:19 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:57:19.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:19 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:57:20.996 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:57:20.997 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:57:21.026 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:57:21.026 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:57:22.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:21 vm08.local ceph-mon[50506]: pgmap v1097: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:21 vm01.local ceph-mon[48597]: pgmap v1097: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:24.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:23 vm08.local ceph-mon[50506]: pgmap v1098: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:57:24.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:23 vm01.local ceph-mon[48597]: pgmap v1098: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:57:25.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:24 vm01.local ceph-mon[48597]: pgmap v1099: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:25.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:24 vm08.local ceph-mon[50506]: pgmap v1099: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:26.028 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:57:26.028 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:57:26.056 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:57:26.057 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:57:27.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:27 vm08.local ceph-mon[50506]: pgmap v1100: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:57:27.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:27 vm01.local ceph-mon[48597]: pgmap v1100: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:57:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:29 vm08.local ceph-mon[50506]: pgmap v1101: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:29 vm01.local ceph-mon[48597]: pgmap v1101: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:31.059 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:57:31.059 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:57:31.085 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:57:31.086 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:57:32.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:31 vm08.local ceph-mon[50506]: pgmap v1102: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:32.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:31 vm01.local ceph-mon[48597]: pgmap v1102: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:34.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:33 vm08.local ceph-mon[50506]: pgmap v1103: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:57:34.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:33 vm01.local ceph-mon[48597]: pgmap v1103: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:57:35.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:34 vm08.local ceph-mon[50506]: pgmap v1104: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:35.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:34 vm01.local ceph-mon[48597]: pgmap v1104: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:36.087 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:57:36.088 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:57:36.114 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:57:36.115 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:57:37.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:37 vm08.local ceph-mon[50506]: pgmap v1105: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:57:37.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:37 vm01.local ceph-mon[48597]: pgmap v1105: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:57:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:39 vm08.local ceph-mon[50506]: pgmap v1106: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:39.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:39 vm01.local ceph-mon[48597]: pgmap v1106: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:41.116 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:57:41.117 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:57:41.169 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:57:41.169 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:57:42.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:41 vm08.local ceph-mon[50506]: pgmap v1107: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:42.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:41 vm01.local ceph-mon[48597]: pgmap v1107: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:44.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:43 vm08.local ceph-mon[50506]: pgmap v1108: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:57:44.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:43 vm01.local ceph-mon[48597]: pgmap v1108: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:57:46.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:45 vm08.local ceph-mon[50506]: pgmap v1109: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:46.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:45 vm01.local ceph-mon[48597]: pgmap v1109: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:46.171 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:57:46.172 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:57:46.202 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:57:46.203 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:57:47.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:46 vm08.local ceph-mon[50506]: pgmap v1110: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:57:47.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:46 vm01.local ceph-mon[48597]: pgmap v1110: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:57:49.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:49 vm08.local ceph-mon[50506]: pgmap v1111: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:49.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:49 vm01.local ceph-mon[48597]: pgmap v1111: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:51.204 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:57:51.205 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:57:51.231 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:57:51.231 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:57:51.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:51 vm08.local ceph-mon[50506]: pgmap v1112: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:51.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:51 vm01.local ceph-mon[48597]: pgmap v1112: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:53.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:53 vm08.local ceph-mon[50506]: pgmap v1113: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:57:53.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:53 vm01.local ceph-mon[48597]: pgmap v1113: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:57:55.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:55 vm08.local ceph-mon[50506]: pgmap v1114: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:55.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:55 vm01.local ceph-mon[48597]: pgmap v1114: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:57:56.233 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:57:56.233 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:57:56.258 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:57:56.259 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:57:57.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:57 vm08.local ceph-mon[50506]: pgmap v1115: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:57:57.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:57 vm01.local ceph-mon[48597]: pgmap v1115: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:58:00.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:57:59 vm08.local ceph-mon[50506]: pgmap v1116: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:00.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:57:59 vm01.local ceph-mon[48597]: pgmap v1116: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:01.260 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:58:01.261 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:58:01.446 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:58:01.447 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:58:02.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:01 vm08.local ceph-mon[50506]: pgmap v1117: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:02.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:01 vm01.local ceph-mon[48597]: pgmap v1117: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:04.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:03 vm08.local ceph-mon[50506]: pgmap v1118: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:58:04.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:03 vm01.local ceph-mon[48597]: pgmap v1118: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:58:06.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:05 vm08.local ceph-mon[50506]: pgmap v1119: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:06.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:05 vm01.local ceph-mon[48597]: pgmap v1119: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:06.448 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:58:06.449 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:58:06.473 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:58:06.474 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:58:08.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:07 vm08.local ceph-mon[50506]: pgmap v1120: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:58:08.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:07 vm01.local ceph-mon[48597]: pgmap v1120: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:58:10.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:09 vm08.local ceph-mon[50506]: pgmap v1121: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:10.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:09 vm01.local ceph-mon[48597]: pgmap v1121: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:58:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:58:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:58:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:58:11.475 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:58:11.476 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:58:11.502 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:58:11.503 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:58:12.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:11 vm08.local ceph-mon[50506]: pgmap v1122: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:12.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:11 vm01.local ceph-mon[48597]: pgmap v1122: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:13.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:12 vm08.local ceph-mon[50506]: pgmap v1123: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:58:13.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:12 vm01.local ceph-mon[48597]: pgmap v1123: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:58:15.797 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:15 vm08.local ceph-mon[50506]: pgmap v1124: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:15.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:15 vm01.local ceph-mon[48597]: pgmap v1124: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:16.504 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:58:16.505 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:58:16.532 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:58:16.533 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:58:18.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:17 vm08.local ceph-mon[50506]: pgmap v1125: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:58:18.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:17 vm01.local ceph-mon[48597]: pgmap v1125: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:58:19.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:18 vm08.local ceph-mon[50506]: pgmap v1126: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:19.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:18 vm01.local ceph-mon[48597]: pgmap v1126: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:19 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:58:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:19 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:58:20.095 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:19 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:58:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:19 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:58:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:19 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:58:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:19 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:58:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:20 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:58:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:20 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:58:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:20 vm08.local ceph-mon[50506]: pgmap v1127: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:20 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:58:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:20 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:58:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:20 vm01.local ceph-mon[48597]: pgmap v1127: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:21.534 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:58:21.535 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:58:21.559 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:58:21.560 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:58:23.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:23 vm08.local ceph-mon[50506]: pgmap v1128: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:58:23.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:23 vm01.local ceph-mon[48597]: pgmap v1128: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:58:25.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:25 vm08.local ceph-mon[50506]: pgmap v1129: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:25 vm01.local ceph-mon[48597]: pgmap v1129: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:26.561 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:58:26.561 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:58:26.588 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:58:26.589 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:58:27.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:27 vm08.local ceph-mon[50506]: pgmap v1130: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:58:27.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:27 vm01.local ceph-mon[48597]: pgmap v1130: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:58:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:29 vm08.local ceph-mon[50506]: pgmap v1131: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:29 vm01.local ceph-mon[48597]: pgmap v1131: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:31.590 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:58:31.591 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:58:31.822 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:58:31.823 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:58:32.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:31 vm01.local ceph-mon[48597]: pgmap v1132: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:32.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:31 vm08.local ceph-mon[50506]: pgmap v1132: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:33.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:32 vm01.local ceph-mon[48597]: pgmap v1133: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:58:33.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:32 vm08.local ceph-mon[50506]: pgmap v1133: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:58:35.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:35 vm08.local ceph-mon[50506]: pgmap v1134: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:35.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:35 vm01.local ceph-mon[48597]: pgmap v1134: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:36.824 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:58:36.825 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:58:36.850 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:58:36.851 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:58:37.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:37 vm01.local ceph-mon[48597]: pgmap v1135: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:58:38.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:37 vm08.local ceph-mon[50506]: pgmap v1135: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:58:40.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:39 vm08.local ceph-mon[50506]: pgmap v1136: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:39 vm01.local ceph-mon[48597]: pgmap v1136: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:41.852 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:58:41.852 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:58:41.877 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:58:41.877 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:58:42.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:41 vm08.local ceph-mon[50506]: pgmap v1137: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:42.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:41 vm01.local ceph-mon[48597]: pgmap v1137: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:44.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:43 vm08.local ceph-mon[50506]: pgmap v1138: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:58:44.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:43 vm01.local ceph-mon[48597]: pgmap v1138: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:58:46.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:45 vm08.local ceph-mon[50506]: pgmap v1139: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:46.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:45 vm01.local ceph-mon[48597]: pgmap v1139: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:46.878 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:58:46.879 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:58:46.906 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:58:46.906 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:58:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:48 vm08.local ceph-mon[50506]: pgmap v1140: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:58:48.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:47 vm01.local ceph-mon[48597]: pgmap v1140: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:58:49.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:48 vm01.local ceph-mon[48597]: pgmap v1141: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:48 vm08.local ceph-mon[50506]: pgmap v1141: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:51.908 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:58:51.908 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:58:51.935 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:58:51.936 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:58:52.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:51 vm08.local ceph-mon[50506]: pgmap v1142: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:52.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:51 vm01.local ceph-mon[48597]: pgmap v1142: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:54.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:53 vm08.local ceph-mon[50506]: pgmap v1143: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:58:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:53 vm01.local ceph-mon[48597]: pgmap v1143: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:58:55.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:55 vm08.local ceph-mon[50506]: pgmap v1144: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:55 vm01.local ceph-mon[48597]: pgmap v1144: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:58:56.937 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:58:56.938 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:58:57.001 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:58:57.002 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:58:58.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:57 vm08.local ceph-mon[50506]: pgmap v1145: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:58:58.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:57 vm01.local ceph-mon[48597]: pgmap v1145: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:59:00.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:58:59 vm08.local ceph-mon[50506]: pgmap v1146: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:00.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:58:59 vm01.local ceph-mon[48597]: pgmap v1146: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:02.003 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:59:02.004 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:59:02.030 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:59:02.031 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:59:02.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:01 vm08.local ceph-mon[50506]: pgmap v1147: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:02.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:01 vm01.local ceph-mon[48597]: pgmap v1147: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:04.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:03 vm08.local ceph-mon[50506]: pgmap v1148: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:59:04.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:03 vm01.local ceph-mon[48597]: pgmap v1148: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:59:06.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:05 vm08.local ceph-mon[50506]: pgmap v1149: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:06.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:05 vm01.local ceph-mon[48597]: pgmap v1149: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:07.032 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:59:07.033 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:59:07.060 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:59:07.060 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:59:08.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:07 vm08.local ceph-mon[50506]: pgmap v1150: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:59:08.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:07 vm01.local ceph-mon[48597]: pgmap v1150: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:59:10.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:09 vm08.local ceph-mon[50506]: pgmap v1151: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:10.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:09 vm01.local ceph-mon[48597]: pgmap v1151: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:59:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:59:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:59:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T10:59:12.062 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:59:12.063 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:59:12.089 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:59:12.090 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:59:12.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:11 vm08.local ceph-mon[50506]: pgmap v1152: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:12.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:11 vm01.local ceph-mon[48597]: pgmap v1152: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:14.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:13 vm08.local ceph-mon[50506]: pgmap v1153: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:59:14.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:13 vm01.local ceph-mon[48597]: pgmap v1153: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:59:16.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:15 vm08.local ceph-mon[50506]: pgmap v1154: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:15 vm01.local ceph-mon[48597]: pgmap v1154: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:17.091 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:59:17.092 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:59:17.118 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:59:17.119 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:59:18.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:17 vm08.local ceph-mon[50506]: pgmap v1155: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:59:18.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:17 vm01.local ceph-mon[48597]: pgmap v1155: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:59:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:19 vm08.local ceph-mon[50506]: pgmap v1156: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:19 vm01.local ceph-mon[48597]: pgmap v1156: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:20.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:20 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:59:20.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:20 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:59:20.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:20 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:59:20.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:20 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:59:20.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:20 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:59:21.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:20 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:59:21.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:20 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:59:21.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:20 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:59:21.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:20 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:59:21.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:20 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T10:59:22.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:21 vm08.local ceph-mon[50506]: pgmap v1157: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:21 vm01.local ceph-mon[48597]: pgmap v1157: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:22.120 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:59:22.120 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:59:22.146 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:59:22.147 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:59:24.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:23 vm08.local ceph-mon[50506]: pgmap v1158: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:59:24.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:23 vm01.local ceph-mon[48597]: pgmap v1158: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:59:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:25 vm08.local ceph-mon[50506]: pgmap v1159: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:25 vm01.local ceph-mon[48597]: pgmap v1159: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:27.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:26 vm08.local ceph-mon[50506]: pgmap v1160: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:59:27.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:26 vm01.local ceph-mon[48597]: pgmap v1160: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:59:27.148 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:59:27.149 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:59:27.174 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:59:27.175 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:59:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:29 vm08.local ceph-mon[50506]: pgmap v1161: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:29 vm01.local ceph-mon[48597]: pgmap v1161: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:32.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:31 vm08.local ceph-mon[50506]: pgmap v1162: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:32.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:31 vm01.local ceph-mon[48597]: pgmap v1162: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:32.176 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:59:32.176 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:59:32.203 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:59:32.204 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:59:34.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:33 vm08.local ceph-mon[50506]: pgmap v1163: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:59:34.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:33 vm01.local ceph-mon[48597]: pgmap v1163: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:59:36.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:35 vm08.local ceph-mon[50506]: pgmap v1164: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:35 vm01.local ceph-mon[48597]: pgmap v1164: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:37.205 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:59:37.205 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:59:37.232 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:59:37.233 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:59:38.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:37 vm08.local ceph-mon[50506]: pgmap v1165: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:59:38.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:37 vm01.local ceph-mon[48597]: pgmap v1165: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:59:40.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:39 vm08.local ceph-mon[50506]: pgmap v1166: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:39 vm01.local ceph-mon[48597]: pgmap v1166: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:42.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:41 vm08.local ceph-mon[50506]: pgmap v1167: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:42.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:41 vm01.local ceph-mon[48597]: pgmap v1167: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:42.234 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:59:42.235 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:59:42.260 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:59:42.261 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:59:44.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:43 vm08.local ceph-mon[50506]: pgmap v1168: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:59:44.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:43 vm01.local ceph-mon[48597]: pgmap v1168: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:59:45.921 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:45 vm01.local ceph-mon[48597]: pgmap v1169: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:46.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:45 vm08.local ceph-mon[50506]: pgmap v1169: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:47.263 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:59:47.263 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:59:47.290 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:59:47.290 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:59:48.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:47 vm08.local ceph-mon[50506]: pgmap v1170: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:59:48.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:47 vm01.local ceph-mon[48597]: pgmap v1170: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:59:50.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:49 vm08.local ceph-mon[50506]: pgmap v1171: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:50.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:49 vm01.local ceph-mon[48597]: pgmap v1171: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:52.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:51 vm08.local ceph-mon[50506]: pgmap v1172: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:52.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:51 vm01.local ceph-mon[48597]: pgmap v1172: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:52.292 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:59:52.292 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:59:52.374 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:59:52.375 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:59:54.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:53 vm08.local ceph-mon[50506]: pgmap v1173: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:59:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:53 vm01.local ceph-mon[48597]: pgmap v1173: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:59:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:55 vm08.local ceph-mon[50506]: pgmap v1174: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:56.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:55 vm01.local ceph-mon[48597]: pgmap v1174: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:57.376 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T10:59:57.377 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T10:59:57.404 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T10:59:57.404 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T10:59:58.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:57 vm01.local ceph-mon[48597]: pgmap v1175: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:59:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:57 vm08.local ceph-mon[50506]: pgmap v1175: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T10:59:59.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 10:59:58 vm08.local ceph-mon[50506]: pgmap v1176: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T10:59:59.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 10:59:58 vm01.local ceph-mon[48597]: pgmap v1176: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:00.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:00 vm08.local ceph-mon[50506]: overall HEALTH_OK 2026-03-10T11:00:00.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:00 vm01.local ceph-mon[48597]: overall HEALTH_OK 2026-03-10T11:00:01.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:01 vm01.local ceph-mon[48597]: pgmap v1177: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:00:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:01 vm08.local ceph-mon[50506]: pgmap v1177: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:00:02.406 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:00:02.407 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:00:02.440 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:00:02.440 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:00:03.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:03 vm08.local ceph-mon[50506]: pgmap v1178: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:00:03.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:03 vm01.local ceph-mon[48597]: pgmap v1178: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:00:06.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:05 vm08.local ceph-mon[50506]: pgmap v1179: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:06.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:05 vm01.local ceph-mon[48597]: pgmap v1179: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:07.442 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:00:07.443 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:00:07.728 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:00:07.728 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:00:08.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:07 vm01.local ceph-mon[48597]: pgmap v1180: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:00:08.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:07 vm08.local ceph-mon[50506]: pgmap v1180: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:00:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:09 vm08.local ceph-mon[50506]: pgmap v1181: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:09.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:09 vm01.local ceph-mon[48597]: pgmap v1181: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:10.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:00:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:00:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:00:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:11 vm01.local ceph-mon[48597]: pgmap v1182: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:00:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:00:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:11 vm08.local ceph-mon[50506]: pgmap v1182: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:00:12.730 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:00:12.730 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:00:12.755 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:00:12.756 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:00:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:13 vm08.local ceph-mon[50506]: pgmap v1183: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:13 vm01.local ceph-mon[48597]: pgmap v1183: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:15.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:15 vm08.local ceph-mon[50506]: pgmap v1184: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:15.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:15 vm01.local ceph-mon[48597]: pgmap v1184: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:17.758 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:00:17.758 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:00:17.784 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:00:17.784 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:00:18.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:17 vm08.local ceph-mon[50506]: pgmap v1185: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:00:18.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:17 vm01.local ceph-mon[48597]: pgmap v1185: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:00:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:19 vm08.local ceph-mon[50506]: pgmap v1186: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:19 vm01.local ceph-mon[48597]: pgmap v1186: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:20.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:20 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:00:20.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:20 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:00:20.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:20 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:00:20.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:20 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:00:20.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:20 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:00:21.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:20 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:00:21.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:20 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:00:21.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:20 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:00:21.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:20 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:00:21.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:20 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:00:22.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:21 vm08.local ceph-mon[50506]: pgmap v1187: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:00:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:21 vm01.local ceph-mon[48597]: pgmap v1187: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:00:22.786 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:00:22.786 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:00:22.813 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:00:22.813 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:00:24.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:23 vm08.local ceph-mon[50506]: pgmap v1188: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:24.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:23 vm01.local ceph-mon[48597]: pgmap v1188: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:25 vm01.local ceph-mon[48597]: pgmap v1189: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:26.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:25 vm08.local ceph-mon[50506]: pgmap v1189: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:27.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:26 vm01.local ceph-mon[48597]: pgmap v1190: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:00:27.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:26 vm08.local ceph-mon[50506]: pgmap v1190: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:00:27.814 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:00:27.815 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:00:27.843 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:00:27.844 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:00:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:29 vm08.local ceph-mon[50506]: pgmap v1191: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:29 vm01.local ceph-mon[48597]: pgmap v1191: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:31.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:31 vm08.local ceph-mon[50506]: pgmap v1192: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:00:31.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:31 vm01.local ceph-mon[48597]: pgmap v1192: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:00:32.845 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:00:32.845 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:00:32.872 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:00:32.872 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:00:33.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:33 vm08.local ceph-mon[50506]: pgmap v1193: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:33.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:33 vm01.local ceph-mon[48597]: pgmap v1193: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:36.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:35 vm08.local ceph-mon[50506]: pgmap v1194: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:35 vm01.local ceph-mon[48597]: pgmap v1194: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:37.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:36 vm08.local ceph-mon[50506]: pgmap v1195: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:00:37.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:36 vm01.local ceph-mon[48597]: pgmap v1195: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:00:37.873 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:00:37.874 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:00:37.901 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:00:37.902 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:00:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:39 vm08.local ceph-mon[50506]: pgmap v1196: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:39.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:39 vm01.local ceph-mon[48597]: pgmap v1196: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:41.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:41 vm01.local ceph-mon[48597]: pgmap v1197: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:00:42.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:41 vm08.local ceph-mon[50506]: pgmap v1197: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:00:42.903 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:00:42.904 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:00:42.986 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:00:42.987 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:00:44.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:43 vm08.local ceph-mon[50506]: pgmap v1198: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:44.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:43 vm01.local ceph-mon[48597]: pgmap v1198: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:46.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:45 vm08.local ceph-mon[50506]: pgmap v1199: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:46.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:45 vm01.local ceph-mon[48597]: pgmap v1199: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:47.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:46 vm08.local ceph-mon[50506]: pgmap v1200: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:00:47.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:46 vm01.local ceph-mon[48597]: pgmap v1200: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:00:47.988 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:00:47.989 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:00:48.017 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:00:48.018 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:00:49.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:49 vm08.local ceph-mon[50506]: pgmap v1201: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:49.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:49 vm01.local ceph-mon[48597]: pgmap v1201: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:51.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:51 vm08.local ceph-mon[50506]: pgmap v1202: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:00:51.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:51 vm01.local ceph-mon[48597]: pgmap v1202: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:00:53.019 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:00:53.019 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:00:53.046 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:00:53.046 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:00:53.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:53 vm08.local ceph-mon[50506]: pgmap v1203: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:53.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:53 vm01.local ceph-mon[48597]: pgmap v1203: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:55 vm08.local ceph-mon[50506]: pgmap v1204: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:56.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:55 vm01.local ceph-mon[48597]: pgmap v1204: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:57.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:56 vm08.local ceph-mon[50506]: pgmap v1205: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:00:57.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:56 vm01.local ceph-mon[48597]: pgmap v1205: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:00:58.047 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:00:58.048 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:00:58.075 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:00:58.075 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:00:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:00:59 vm08.local ceph-mon[50506]: pgmap v1206: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:00:59.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:00:59 vm01.local ceph-mon[48597]: pgmap v1206: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:01 vm08.local ceph-mon[50506]: pgmap v1207: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:01:01.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:01 vm01.local ceph-mon[48597]: pgmap v1207: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:01:03.076 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:01:03.077 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:01:03.154 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:01:03.155 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:01:03.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:03 vm08.local ceph-mon[50506]: pgmap v1208: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:03.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:03 vm01.local ceph-mon[48597]: pgmap v1208: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:05 vm08.local ceph-mon[50506]: pgmap v1209: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:05 vm01.local ceph-mon[48597]: pgmap v1209: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:07.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:06 vm08.local ceph-mon[50506]: pgmap v1210: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:01:07.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:06 vm01.local ceph-mon[48597]: pgmap v1210: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:01:08.156 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:01:08.157 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:01:08.183 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:01:08.184 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:01:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:09 vm08.local ceph-mon[50506]: pgmap v1211: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:09 vm01.local ceph-mon[48597]: pgmap v1211: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:01:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:01:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:01:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:01:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:11 vm08.local ceph-mon[50506]: pgmap v1212: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:01:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:11 vm01.local ceph-mon[48597]: pgmap v1212: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:01:13.185 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:01:13.186 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:01:13.214 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:01:13.215 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:01:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:13 vm08.local ceph-mon[50506]: pgmap v1213: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:13 vm01.local ceph-mon[48597]: pgmap v1213: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:15.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:15 vm01.local ceph-mon[48597]: pgmap v1214: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:16.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:15 vm08.local ceph-mon[50506]: pgmap v1214: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:17.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:16 vm08.local ceph-mon[50506]: pgmap v1215: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:01:17.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:16 vm01.local ceph-mon[48597]: pgmap v1215: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:01:18.216 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:01:18.216 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:01:18.242 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:01:18.243 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:01:19.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:19 vm08.local ceph-mon[50506]: pgmap v1216: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:19.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:19 vm01.local ceph-mon[48597]: pgmap v1216: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:20.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:20 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:01:20.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:20 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:01:20.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:20 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:01:20.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:20 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:01:20.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:20 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:01:20.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:20 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:01:22.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:21 vm08.local ceph-mon[50506]: pgmap v1217: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:01:22.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:21 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:01:22.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:21 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:01:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:21 vm01.local ceph-mon[48597]: pgmap v1217: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:01:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:21 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:01:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:21 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:01:23.244 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:01:23.245 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:01:23.273 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:01:23.273 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:01:24.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:23 vm08.local ceph-mon[50506]: pgmap v1218: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:24.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:23 vm01.local ceph-mon[48597]: pgmap v1218: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:25 vm08.local ceph-mon[50506]: pgmap v1219: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:25 vm01.local ceph-mon[48597]: pgmap v1219: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:27.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:26 vm08.local ceph-mon[50506]: pgmap v1220: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:01:27.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:26 vm01.local ceph-mon[48597]: pgmap v1220: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:01:28.275 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:01:28.275 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:01:28.300 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:01:28.301 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:01:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:29 vm08.local ceph-mon[50506]: pgmap v1221: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:29 vm01.local ceph-mon[48597]: pgmap v1221: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:31.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:31 vm08.local ceph-mon[50506]: pgmap v1222: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:01:31.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:31 vm01.local ceph-mon[48597]: pgmap v1222: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:01:33.302 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:01:33.303 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:01:33.331 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:01:33.332 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:01:33.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:33 vm08.local ceph-mon[50506]: pgmap v1223: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:33.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:33 vm01.local ceph-mon[48597]: pgmap v1223: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:35.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:35 vm01.local ceph-mon[48597]: pgmap v1224: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:36.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:35 vm08.local ceph-mon[50506]: pgmap v1224: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:37.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:36 vm08.local ceph-mon[50506]: pgmap v1225: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:01:37.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:36 vm01.local ceph-mon[48597]: pgmap v1225: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:01:38.333 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:01:38.334 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:01:38.361 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:01:38.361 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:01:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:39 vm08.local ceph-mon[50506]: pgmap v1226: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:39.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:39 vm01.local ceph-mon[48597]: pgmap v1226: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:41.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:41 vm08.local ceph-mon[50506]: pgmap v1227: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:01:41.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:41 vm01.local ceph-mon[48597]: pgmap v1227: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:01:43.363 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:01:43.363 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:01:43.389 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:01:43.389 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:01:43.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:43 vm08.local ceph-mon[50506]: pgmap v1228: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:43.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:43 vm01.local ceph-mon[48597]: pgmap v1228: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:45.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:45 vm01.local ceph-mon[48597]: pgmap v1229: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:46.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:45 vm08.local ceph-mon[50506]: pgmap v1229: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:47.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:46 vm08.local ceph-mon[50506]: pgmap v1230: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:01:47.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:46 vm01.local ceph-mon[48597]: pgmap v1230: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:01:48.391 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:01:48.392 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:01:48.417 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:01:48.418 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:01:49.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:49 vm08.local ceph-mon[50506]: pgmap v1231: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:49.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:49 vm01.local ceph-mon[48597]: pgmap v1231: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:51.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:51 vm08.local ceph-mon[50506]: pgmap v1232: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:01:51.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:51 vm01.local ceph-mon[48597]: pgmap v1232: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:01:53.419 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:01:53.420 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:01:53.445 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:01:53.445 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:01:53.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:53 vm08.local ceph-mon[50506]: pgmap v1233: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:53.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:53 vm01.local ceph-mon[48597]: pgmap v1233: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:55.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:55 vm01.local ceph-mon[48597]: pgmap v1234: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:55 vm08.local ceph-mon[50506]: pgmap v1234: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:57.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:56 vm01.local ceph-mon[48597]: pgmap v1235: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:01:57.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:56 vm08.local ceph-mon[50506]: pgmap v1235: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:01:58.447 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:01:58.447 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:01:58.473 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:01:58.473 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:01:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:01:59 vm08.local ceph-mon[50506]: pgmap v1236: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:01:59.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:01:59 vm01.local ceph-mon[48597]: pgmap v1236: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:01 vm08.local ceph-mon[50506]: pgmap v1237: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:02:01.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:01 vm01.local ceph-mon[48597]: pgmap v1237: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:02:03.474 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:02:03.475 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:02:03.500 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:02:03.501 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:02:03.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:03 vm01.local ceph-mon[48597]: pgmap v1238: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:04.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:03 vm08.local ceph-mon[50506]: pgmap v1238: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:05 vm01.local ceph-mon[48597]: pgmap v1239: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:06.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:05 vm08.local ceph-mon[50506]: pgmap v1239: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:07.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:06 vm01.local ceph-mon[48597]: pgmap v1240: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:02:07.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:06 vm08.local ceph-mon[50506]: pgmap v1240: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:02:08.502 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:02:08.503 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:02:08.528 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:02:08.529 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:02:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:09 vm08.local ceph-mon[50506]: pgmap v1241: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:09 vm01.local ceph-mon[48597]: pgmap v1241: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:02:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:02:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:02:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:02:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:11 vm01.local ceph-mon[48597]: pgmap v1242: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:02:12.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:11 vm08.local ceph-mon[50506]: pgmap v1242: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:02:13.530 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:02:13.531 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:02:13.556 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:02:13.557 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:02:13.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:13 vm01.local ceph-mon[48597]: pgmap v1243: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:14.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:13 vm08.local ceph-mon[50506]: pgmap v1243: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:15.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:15 vm01.local ceph-mon[48597]: pgmap v1244: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:15.896 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:15 vm08.local ceph-mon[50506]: pgmap v1244: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:17.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:16 vm01.local ceph-mon[48597]: pgmap v1245: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:02:17.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:16 vm08.local ceph-mon[50506]: pgmap v1245: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:02:18.558 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:02:18.558 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:02:18.584 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:02:18.584 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:02:19.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:19 vm08.local ceph-mon[50506]: pgmap v1246: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:19.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:19 vm01.local ceph-mon[48597]: pgmap v1246: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:21.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:21 vm08.local ceph-mon[50506]: pgmap v1247: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:02:21.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:21 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:02:21.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:21 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:02:21.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:21 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:02:21.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:21 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:02:21.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:21 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:02:21.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:21 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:02:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:21 vm01.local ceph-mon[48597]: pgmap v1247: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:02:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:21 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:02:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:21 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:02:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:21 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:02:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:21 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:02:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:21 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:02:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:21 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:02:23.585 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:02:23.586 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:02:23.612 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:02:23.613 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:02:23.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:23 vm01.local ceph-mon[48597]: pgmap v1248: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:24.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:23 vm08.local ceph-mon[50506]: pgmap v1248: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:25 vm01.local ceph-mon[48597]: pgmap v1249: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:25 vm08.local ceph-mon[50506]: pgmap v1249: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:27.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:26 vm01.local ceph-mon[48597]: pgmap v1250: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:02:27.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:26 vm08.local ceph-mon[50506]: pgmap v1250: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:02:28.614 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:02:28.614 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:02:28.639 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:02:28.640 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:02:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:29 vm08.local ceph-mon[50506]: pgmap v1251: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:29 vm01.local ceph-mon[48597]: pgmap v1251: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:31.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:31 vm08.local ceph-mon[50506]: pgmap v1252: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:02:31.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:31 vm01.local ceph-mon[48597]: pgmap v1252: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:02:33.641 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:02:33.642 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:02:33.667 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:02:33.667 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:02:33.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:33 vm01.local ceph-mon[48597]: pgmap v1253: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:34.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:33 vm08.local ceph-mon[50506]: pgmap v1253: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:35.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:35 vm01.local ceph-mon[48597]: pgmap v1254: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:36.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:35 vm08.local ceph-mon[50506]: pgmap v1254: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:37.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:36 vm08.local ceph-mon[50506]: pgmap v1255: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:02:37.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:36 vm01.local ceph-mon[48597]: pgmap v1255: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:02:38.669 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:02:38.669 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:02:38.694 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:02:38.694 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:02:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:39 vm08.local ceph-mon[50506]: pgmap v1256: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:39.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:39 vm01.local ceph-mon[48597]: pgmap v1256: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:41.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:41 vm01.local ceph-mon[48597]: pgmap v1257: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:02:42.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:41 vm08.local ceph-mon[50506]: pgmap v1257: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:02:43.696 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:02:43.696 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:02:43.722 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:02:43.722 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:02:43.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:43 vm01.local ceph-mon[48597]: pgmap v1258: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:44.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:43 vm08.local ceph-mon[50506]: pgmap v1258: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:45.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:45 vm01.local ceph-mon[48597]: pgmap v1259: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:45.897 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:45 vm08.local ceph-mon[50506]: pgmap v1259: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:47.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:46 vm01.local ceph-mon[48597]: pgmap v1260: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:02:47.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:46 vm08.local ceph-mon[50506]: pgmap v1260: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:02:48.724 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:02:48.724 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:02:48.750 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:02:48.751 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:02:49.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:49 vm08.local ceph-mon[50506]: pgmap v1261: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:49.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:49 vm01.local ceph-mon[48597]: pgmap v1261: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:52.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:51 vm08.local ceph-mon[50506]: pgmap v1262: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:02:52.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:51 vm01.local ceph-mon[48597]: pgmap v1262: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:02:53.752 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:02:53.753 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:02:53.779 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:02:53.780 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:02:54.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:53 vm08.local ceph-mon[50506]: pgmap v1263: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:53 vm01.local ceph-mon[48597]: pgmap v1263: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:55 vm08.local ceph-mon[50506]: pgmap v1264: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:56.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:55 vm01.local ceph-mon[48597]: pgmap v1264: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:02:57.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:56 vm01.local ceph-mon[48597]: pgmap v1265: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:02:57.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:56 vm08.local ceph-mon[50506]: pgmap v1265: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:02:58.781 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:02:58.782 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:02:58.807 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:02:58.808 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:02:59.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:02:59 vm01.local ceph-mon[48597]: pgmap v1266: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:00.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:02:59 vm08.local ceph-mon[50506]: pgmap v1266: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:01.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:01 vm01.local ceph-mon[48597]: pgmap v1267: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:03:02.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:01 vm08.local ceph-mon[50506]: pgmap v1267: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:03:03.809 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:03:03.810 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:03:03.835 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:03:03.836 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:03:03.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:03 vm01.local ceph-mon[48597]: pgmap v1268: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:04.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:03 vm08.local ceph-mon[50506]: pgmap v1268: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:06.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:05 vm08.local ceph-mon[50506]: pgmap v1269: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:06.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:05 vm01.local ceph-mon[48597]: pgmap v1269: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:07.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:06 vm01.local ceph-mon[48597]: pgmap v1270: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:03:07.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:06 vm08.local ceph-mon[50506]: pgmap v1270: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:03:08.837 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:03:08.838 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:03:08.878 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:03:08.879 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:03:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:09 vm01.local ceph-mon[48597]: pgmap v1271: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:10.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:09 vm08.local ceph-mon[50506]: pgmap v1271: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:03:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:03:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:03:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:03:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:11 vm01.local ceph-mon[48597]: pgmap v1272: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:03:12.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:11 vm08.local ceph-mon[50506]: pgmap v1272: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:03:13.880 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:03:13.881 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:03:13.906 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:03:13.907 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:03:14.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:13 vm08.local ceph-mon[50506]: pgmap v1273: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:14.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:13 vm01.local ceph-mon[48597]: pgmap v1273: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:15.897 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:15 vm08.local ceph-mon[50506]: pgmap v1274: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:15.921 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:15 vm01.local ceph-mon[48597]: pgmap v1274: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:17.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:16 vm08.local ceph-mon[50506]: pgmap v1275: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:03:17.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:16 vm01.local ceph-mon[48597]: pgmap v1275: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:03:18.908 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:03:18.909 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:03:18.969 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:03:18.969 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:03:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:19 vm08.local ceph-mon[50506]: pgmap v1276: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:19 vm01.local ceph-mon[48597]: pgmap v1276: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:21.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:21 vm08.local ceph-mon[50506]: pgmap v1277: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:03:21.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:21 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:03:21.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:21 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:03:21.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:21 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:03:21.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:21 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:03:21.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:21 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:03:21.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:21 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:03:21.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:21 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:03:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:21 vm01.local ceph-mon[48597]: pgmap v1277: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:03:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:21 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:03:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:21 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:03:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:21 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:03:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:21 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:03:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:21 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:03:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:21 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:03:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:21 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:03:23.971 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:03:23.972 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:03:24.005 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:03:24.005 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:03:24.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:23 vm08.local ceph-mon[50506]: pgmap v1278: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:24.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:23 vm01.local ceph-mon[48597]: pgmap v1278: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:25 vm08.local ceph-mon[50506]: pgmap v1279: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:25 vm01.local ceph-mon[48597]: pgmap v1279: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:27.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:26 vm08.local ceph-mon[50506]: pgmap v1280: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:03:27.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:26 vm01.local ceph-mon[48597]: pgmap v1280: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:03:29.007 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:03:29.007 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:03:29.038 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:03:29.039 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:03:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:29 vm01.local ceph-mon[48597]: pgmap v1281: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:29 vm08.local ceph-mon[50506]: pgmap v1281: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:31.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:31 vm01.local ceph-mon[48597]: pgmap v1282: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:03:32.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:31 vm08.local ceph-mon[50506]: pgmap v1282: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:03:33.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:33 vm01.local ceph-mon[48597]: pgmap v1283: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:34.040 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:03:34.041 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:03:34.066 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:03:34.067 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:03:34.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:33 vm08.local ceph-mon[50506]: pgmap v1283: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:35.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:35 vm01.local ceph-mon[48597]: pgmap v1284: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:36.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:35 vm08.local ceph-mon[50506]: pgmap v1284: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:37.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:37 vm08.local ceph-mon[50506]: pgmap v1285: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:03:37.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:37 vm01.local ceph-mon[48597]: pgmap v1285: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:03:39.068 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:03:39.069 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:03:39.094 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:03:39.095 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:03:40.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:39 vm08.local ceph-mon[50506]: pgmap v1286: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:39 vm01.local ceph-mon[48597]: pgmap v1286: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:42.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:41 vm08.local ceph-mon[50506]: pgmap v1287: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:03:42.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:41 vm01.local ceph-mon[48597]: pgmap v1287: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:03:44.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:43 vm08.local ceph-mon[50506]: pgmap v1288: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:44.096 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:03:44.097 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:03:44.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:43 vm01.local ceph-mon[48597]: pgmap v1288: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:44.123 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:03:44.123 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:03:45.897 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:45 vm08.local ceph-mon[50506]: pgmap v1289: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:45.919 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:45 vm01.local ceph-mon[48597]: pgmap v1289: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:47.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:46 vm08.local ceph-mon[50506]: pgmap v1290: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:03:47.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:46 vm01.local ceph-mon[48597]: pgmap v1290: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:03:49.125 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:03:49.125 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:03:49.152 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:03:49.153 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:03:49.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:49 vm01.local ceph-mon[48597]: pgmap v1291: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:50.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:49 vm08.local ceph-mon[50506]: pgmap v1291: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:51.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:51 vm01.local ceph-mon[48597]: pgmap v1292: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:03:52.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:51 vm08.local ceph-mon[50506]: pgmap v1292: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:03:54.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:53 vm08.local ceph-mon[50506]: pgmap v1293: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:53 vm01.local ceph-mon[48597]: pgmap v1293: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:54.154 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:03:54.155 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:03:54.185 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:03:54.186 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:03:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:55 vm08.local ceph-mon[50506]: pgmap v1294: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:56.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:55 vm01.local ceph-mon[48597]: pgmap v1294: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:03:57.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:56 vm08.local ceph-mon[50506]: pgmap v1295: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:03:57.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:56 vm01.local ceph-mon[48597]: pgmap v1295: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:03:59.187 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:03:59.188 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:03:59.214 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:03:59.214 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:04:00.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:03:59 vm08.local ceph-mon[50506]: pgmap v1296: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:00.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:03:59 vm01.local ceph-mon[48597]: pgmap v1296: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:02.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:01 vm08.local ceph-mon[50506]: pgmap v1297: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:04:02.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:01 vm01.local ceph-mon[48597]: pgmap v1297: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:04:04.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:03 vm08.local ceph-mon[50506]: pgmap v1298: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:04.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:03 vm01.local ceph-mon[48597]: pgmap v1298: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:04.215 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:04:04.216 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:04:04.241 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:04:04.242 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:04:06.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:05 vm08.local ceph-mon[50506]: pgmap v1299: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:06.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:05 vm01.local ceph-mon[48597]: pgmap v1299: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:07.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:06 vm08.local ceph-mon[50506]: pgmap v1300: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:04:07.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:06 vm01.local ceph-mon[48597]: pgmap v1300: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:04:09.243 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:04:09.244 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:04:09.269 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:04:09.270 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:04:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:09 vm01.local ceph-mon[48597]: pgmap v1301: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:10.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:09 vm08.local ceph-mon[50506]: pgmap v1301: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:04:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:04:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:04:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:04:12.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:11 vm08.local ceph-mon[50506]: pgmap v1302: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:04:12.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:11 vm01.local ceph-mon[48597]: pgmap v1302: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:04:14.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:13 vm08.local ceph-mon[50506]: pgmap v1303: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:14.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:13 vm01.local ceph-mon[48597]: pgmap v1303: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:14.271 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:04:14.271 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:04:14.297 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:04:14.297 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:04:16.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:15 vm08.local ceph-mon[50506]: pgmap v1304: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:15 vm01.local ceph-mon[48597]: pgmap v1304: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:17.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:16 vm08.local ceph-mon[50506]: pgmap v1305: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:04:17.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:16 vm01.local ceph-mon[48597]: pgmap v1305: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:04:19.299 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:04:19.299 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:04:19.516 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:04:19.517 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:04:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:19 vm08.local ceph-mon[50506]: pgmap v1306: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:19 vm01.local ceph-mon[48597]: pgmap v1306: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:21.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:21 vm08.local ceph-mon[50506]: pgmap v1307: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:04:21.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:21 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:04:21.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:21 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:04:21.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:21 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:04:21.869 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:21 vm01.local ceph-mon[48597]: pgmap v1307: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:04:21.869 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:21 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:04:21.869 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:21 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:04:21.869 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:21 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:04:23.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:23 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:04:23.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:23 vm01.local ceph-mon[48597]: pgmap v1308: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:23.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:23 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:04:23.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:23 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:04:23.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:23 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:04:23.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:23 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:04:23.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:23 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:04:23.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:23 vm08.local ceph-mon[50506]: pgmap v1308: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:23.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:23 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:04:23.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:23 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:04:23.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:23 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:04:23.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:23 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:04:24.519 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:04:24.519 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:04:24.557 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:04:24.557 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:04:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:25 vm01.local ceph-mon[48597]: pgmap v1309: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:25 vm08.local ceph-mon[50506]: pgmap v1309: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:27.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:26 vm08.local ceph-mon[50506]: pgmap v1310: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:04:27.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:26 vm01.local ceph-mon[48597]: pgmap v1310: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:04:29.560 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:04:29.560 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:04:29.592 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:04:29.593 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:04:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:29 vm01.local ceph-mon[48597]: pgmap v1311: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:29 vm08.local ceph-mon[50506]: pgmap v1311: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:31.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:31 vm01.local ceph-mon[48597]: pgmap v1312: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:04:32.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:31 vm08.local ceph-mon[50506]: pgmap v1312: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:04:34.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:33 vm08.local ceph-mon[50506]: pgmap v1313: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:34.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:33 vm01.local ceph-mon[48597]: pgmap v1313: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:34.594 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:04:34.594 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:04:34.619 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:04:34.620 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:04:36.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:35 vm08.local ceph-mon[50506]: pgmap v1314: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:35 vm01.local ceph-mon[48597]: pgmap v1314: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:37.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:36 vm08.local ceph-mon[50506]: pgmap v1315: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:04:37.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:36 vm01.local ceph-mon[48597]: pgmap v1315: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:04:39.621 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:04:39.622 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:04:39.648 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:04:39.648 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:04:40.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:39 vm08.local ceph-mon[50506]: pgmap v1316: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:39 vm01.local ceph-mon[48597]: pgmap v1316: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:42.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:41 vm08.local ceph-mon[50506]: pgmap v1317: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:04:42.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:41 vm01.local ceph-mon[48597]: pgmap v1317: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:04:44.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:43 vm08.local ceph-mon[50506]: pgmap v1318: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:44.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:43 vm01.local ceph-mon[48597]: pgmap v1318: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:44.649 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:04:44.650 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:04:44.678 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:04:44.678 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:04:45.920 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:45 vm01.local ceph-mon[48597]: pgmap v1319: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:46.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:45 vm08.local ceph-mon[50506]: pgmap v1319: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:47.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:46 vm08.local ceph-mon[50506]: pgmap v1320: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:04:47.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:46 vm01.local ceph-mon[48597]: pgmap v1320: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:04:49.680 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:04:49.680 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:04:49.911 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:04:49.912 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:04:50.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:49 vm08.local ceph-mon[50506]: pgmap v1321: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:50.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:49 vm01.local ceph-mon[48597]: pgmap v1321: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:51.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:50 vm08.local ceph-mon[50506]: pgmap v1322: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:04:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:50 vm01.local ceph-mon[48597]: pgmap v1322: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:04:53.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:53 vm01.local ceph-mon[48597]: pgmap v1323: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:54.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:53 vm08.local ceph-mon[50506]: pgmap v1323: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:54.913 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:04:54.914 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:04:54.940 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:04:54.940 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:04:56.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:55 vm01.local ceph-mon[48597]: pgmap v1324: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:56.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:55 vm08.local ceph-mon[50506]: pgmap v1324: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:04:57.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:56 vm08.local ceph-mon[50506]: pgmap v1325: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:04:57.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:56 vm01.local ceph-mon[48597]: pgmap v1325: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:04:59.941 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:04:59.942 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:04:59.968 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:04:59.968 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:05:00.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:04:59 vm08.local ceph-mon[50506]: pgmap v1326: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:00.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:04:59 vm01.local ceph-mon[48597]: pgmap v1326: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:02.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:01 vm01.local ceph-mon[48597]: pgmap v1327: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:05:02.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:01 vm08.local ceph-mon[50506]: pgmap v1327: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:05:03.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:02 vm08.local ceph-mon[50506]: pgmap v1328: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:03.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:02 vm01.local ceph-mon[48597]: pgmap v1328: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:04.970 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:05:04.970 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:05:05.069 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:05:05.069 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:05:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:05 vm01.local ceph-mon[48597]: pgmap v1329: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:06.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:05 vm08.local ceph-mon[50506]: pgmap v1329: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:07.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:07 vm08.local ceph-mon[50506]: pgmap v1330: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:05:07.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:07 vm01.local ceph-mon[48597]: pgmap v1330: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:05:10.071 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:05:10.071 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:05:10.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:09 vm08.local ceph-mon[50506]: pgmap v1331: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:10.098 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:05:10.099 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:05:10.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:09 vm01.local ceph-mon[48597]: pgmap v1331: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:05:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:05:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:05:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:05:12.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:11 vm08.local ceph-mon[50506]: pgmap v1332: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:05:12.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:11 vm01.local ceph-mon[48597]: pgmap v1332: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:05:14.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:13 vm08.local ceph-mon[50506]: pgmap v1333: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:14.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:13 vm01.local ceph-mon[48597]: pgmap v1333: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:15.100 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:05:15.101 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:05:15.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:14 vm01.local ceph-mon[48597]: pgmap v1334: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:15.126 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:05:15.127 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:05:15.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:14 vm08.local ceph-mon[50506]: pgmap v1334: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:17.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:17 vm08.local ceph-mon[50506]: pgmap v1335: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:05:17.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:17 vm01.local ceph-mon[48597]: pgmap v1335: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:05:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:19 vm08.local ceph-mon[50506]: pgmap v1336: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:19 vm01.local ceph-mon[48597]: pgmap v1336: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:20.128 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:05:20.129 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:05:20.186 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:05:20.187 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:05:22.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:21 vm08.local ceph-mon[50506]: pgmap v1337: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:05:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:21 vm01.local ceph-mon[48597]: pgmap v1337: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:05:23.648 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:23 vm08.local ceph-mon[50506]: pgmap v1338: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:23.716 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:23 vm01.local ceph-mon[48597]: pgmap v1338: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:23.716 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:23 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:05:23.716 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:23 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:05:23.716 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:23 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:05:24.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:23 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:05:24.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:23 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:05:24.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:23 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:05:25.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:24 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:05:25.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:24 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:05:25.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:24 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:05:25.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:24 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:05:25.188 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:05:25.189 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:05:25.219 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:05:25.219 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:05:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:25 vm08.local ceph-mon[50506]: pgmap v1339: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:25 vm01.local ceph-mon[48597]: pgmap v1339: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:27.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:27 vm08.local ceph-mon[50506]: pgmap v1340: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:05:27.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:27 vm01.local ceph-mon[48597]: pgmap v1340: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:05:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:29 vm08.local ceph-mon[50506]: pgmap v1341: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:29 vm01.local ceph-mon[48597]: pgmap v1341: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:30.222 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:05:30.223 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:05:30.252 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:05:30.253 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:05:32.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:31 vm08.local ceph-mon[50506]: pgmap v1342: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:05:32.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:31 vm01.local ceph-mon[48597]: pgmap v1342: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:05:34.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:33 vm08.local ceph-mon[50506]: pgmap v1343: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:34.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:33 vm01.local ceph-mon[48597]: pgmap v1343: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:35.255 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:05:35.255 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:05:35.281 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:05:35.281 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:05:36.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:35 vm08.local ceph-mon[50506]: pgmap v1344: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:35 vm01.local ceph-mon[48597]: pgmap v1344: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:37.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:37 vm08.local ceph-mon[50506]: pgmap v1345: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:05:37.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:37 vm01.local ceph-mon[48597]: pgmap v1345: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:05:40.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:39 vm08.local ceph-mon[50506]: pgmap v1346: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:39 vm01.local ceph-mon[48597]: pgmap v1346: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:40.282 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:05:40.283 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:05:40.310 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:05:40.310 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:05:42.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:41 vm08.local ceph-mon[50506]: pgmap v1347: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:05:42.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:41 vm01.local ceph-mon[48597]: pgmap v1347: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:05:44.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:43 vm08.local ceph-mon[50506]: pgmap v1348: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:44.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:43 vm01.local ceph-mon[48597]: pgmap v1348: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:45.312 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:05:45.312 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:05:45.372 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:05:45.372 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:05:45.920 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:45 vm01.local ceph-mon[48597]: pgmap v1349: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:46.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:45 vm08.local ceph-mon[50506]: pgmap v1349: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:47.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:47 vm08.local ceph-mon[50506]: pgmap v1350: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:05:47.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:47 vm01.local ceph-mon[48597]: pgmap v1350: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:05:49.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:48 vm01.local ceph-mon[48597]: pgmap v1351: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:48 vm08.local ceph-mon[50506]: pgmap v1351: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:50.373 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:05:50.374 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:05:50.401 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:05:50.401 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:05:52.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:51 vm08.local ceph-mon[50506]: pgmap v1352: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:05:52.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:51 vm01.local ceph-mon[48597]: pgmap v1352: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:05:54.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:53 vm08.local ceph-mon[50506]: pgmap v1353: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:53 vm01.local ceph-mon[48597]: pgmap v1353: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:55.403 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:05:55.403 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:05:55.430 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:05:55.430 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:05:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:55 vm08.local ceph-mon[50506]: pgmap v1354: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:56.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:55 vm01.local ceph-mon[48597]: pgmap v1354: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:05:57.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:57 vm08.local ceph-mon[50506]: pgmap v1355: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:05:57.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:57 vm01.local ceph-mon[48597]: pgmap v1355: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:06:00.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:05:59 vm08.local ceph-mon[50506]: pgmap v1356: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:00.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:05:59 vm01.local ceph-mon[48597]: pgmap v1356: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:00.432 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:06:00.432 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:06:00.459 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:06:00.460 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:06:02.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:01 vm08.local ceph-mon[50506]: pgmap v1357: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:06:02.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:01 vm01.local ceph-mon[48597]: pgmap v1357: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:06:04.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:03 vm08.local ceph-mon[50506]: pgmap v1358: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:04.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:03 vm01.local ceph-mon[48597]: pgmap v1358: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:05.461 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:06:05.461 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:06:05.487 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:06:05.488 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:06:06.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:05 vm08.local ceph-mon[50506]: pgmap v1359: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:06.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:05 vm01.local ceph-mon[48597]: pgmap v1359: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:07.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:07 vm08.local ceph-mon[50506]: pgmap v1360: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:06:07.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:07 vm01.local ceph-mon[48597]: pgmap v1360: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:06:10.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:09 vm08.local ceph-mon[50506]: pgmap v1361: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:10.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:09 vm01.local ceph-mon[48597]: pgmap v1361: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:10.489 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:06:10.490 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:06:10.517 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:06:10.518 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:06:10.994 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:06:10.994 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:06:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:06:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:06:12.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:11 vm08.local ceph-mon[50506]: pgmap v1362: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:06:12.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:11 vm01.local ceph-mon[48597]: pgmap v1362: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:06:14.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:13 vm08.local ceph-mon[50506]: pgmap v1363: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:14.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:13 vm01.local ceph-mon[48597]: pgmap v1363: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:15.520 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:06:15.520 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:06:15.548 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:06:15.549 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:06:15.919 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:15 vm01.local ceph-mon[48597]: pgmap v1364: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:16.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:15 vm08.local ceph-mon[50506]: pgmap v1364: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:17.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:17 vm08.local ceph-mon[50506]: pgmap v1365: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:17.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:17 vm01.local ceph-mon[48597]: pgmap v1365: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:19 vm08.local ceph-mon[50506]: pgmap v1366: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:06:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:19 vm01.local ceph-mon[48597]: pgmap v1366: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:06:20.550 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:06:20.551 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:06:20.577 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:06:20.578 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:06:22.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:21 vm08.local ceph-mon[50506]: pgmap v1367: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:21 vm01.local ceph-mon[48597]: pgmap v1367: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:24.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:23 vm08.local ceph-mon[50506]: pgmap v1368: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:06:24.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:23 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:06:24.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:23 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:06:24.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:23 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:06:24.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:23 vm01.local ceph-mon[48597]: pgmap v1368: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:06:24.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:23 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:06:24.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:23 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:06:24.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:23 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:06:25.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:24 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:06:25.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:24 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:06:25.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:24 vm08.local ceph-mon[50506]: pgmap v1369: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:06:25.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:24 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:06:25.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:24 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:06:25.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:24 vm01.local ceph-mon[48597]: pgmap v1369: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:06:25.579 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:06:25.580 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:06:25.634 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:06:25.635 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:06:27.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:27 vm08.local ceph-mon[50506]: pgmap v1370: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:27.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:27 vm01.local ceph-mon[48597]: pgmap v1370: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:29 vm08.local ceph-mon[50506]: pgmap v1371: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:29 vm01.local ceph-mon[48597]: pgmap v1371: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:30.636 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:06:30.636 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:06:30.664 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:06:30.664 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:06:32.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:31 vm08.local ceph-mon[50506]: pgmap v1372: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:06:32.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:31 vm01.local ceph-mon[48597]: pgmap v1372: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:06:34.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:33 vm08.local ceph-mon[50506]: pgmap v1373: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:34.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:33 vm01.local ceph-mon[48597]: pgmap v1373: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:35.665 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:06:35.666 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:06:35.694 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:06:35.694 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:06:36.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:35 vm08.local ceph-mon[50506]: pgmap v1374: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:35 vm01.local ceph-mon[48597]: pgmap v1374: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:37.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:37 vm08.local ceph-mon[50506]: pgmap v1375: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:06:37.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:37 vm01.local ceph-mon[48597]: pgmap v1375: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:06:40.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:39 vm08.local ceph-mon[50506]: pgmap v1376: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:39 vm01.local ceph-mon[48597]: pgmap v1376: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:40.695 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:06:40.696 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:06:40.721 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:06:40.722 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:06:42.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:41 vm08.local ceph-mon[50506]: pgmap v1377: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:06:42.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:41 vm01.local ceph-mon[48597]: pgmap v1377: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:06:44.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:43 vm08.local ceph-mon[50506]: pgmap v1378: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:44.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:43 vm01.local ceph-mon[48597]: pgmap v1378: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:45.723 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:06:45.724 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:06:45.750 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:06:45.751 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:06:45.920 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:45 vm01.local ceph-mon[48597]: pgmap v1379: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:46.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:45 vm08.local ceph-mon[50506]: pgmap v1379: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:47.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:47 vm01.local ceph-mon[48597]: pgmap v1380: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:06:47.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:47 vm08.local ceph-mon[50506]: pgmap v1380: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:06:50.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:49 vm08.local ceph-mon[50506]: pgmap v1381: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:50.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:49 vm01.local ceph-mon[48597]: pgmap v1381: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:50.752 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:06:50.753 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:06:50.780 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:06:50.780 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:06:52.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:51 vm08.local ceph-mon[50506]: pgmap v1382: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:06:52.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:51 vm01.local ceph-mon[48597]: pgmap v1382: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:06:54.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:53 vm08.local ceph-mon[50506]: pgmap v1383: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:53 vm01.local ceph-mon[48597]: pgmap v1383: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:55.782 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:06:55.783 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:06:55.809 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:06:55.809 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:06:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:55 vm08.local ceph-mon[50506]: pgmap v1384: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:56.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:55 vm01.local ceph-mon[48597]: pgmap v1384: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:06:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:57 vm08.local ceph-mon[50506]: pgmap v1385: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:06:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:57 vm01.local ceph-mon[48597]: pgmap v1385: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:07:00.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:06:59 vm08.local ceph-mon[50506]: pgmap v1386: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:00.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:06:59 vm01.local ceph-mon[48597]: pgmap v1386: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:00.811 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:07:00.812 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:07:00.846 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:07:00.847 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:07:02.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:01 vm08.local ceph-mon[50506]: pgmap v1387: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:07:02.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:01 vm01.local ceph-mon[48597]: pgmap v1387: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:07:04.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:03 vm08.local ceph-mon[50506]: pgmap v1388: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:04.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:03 vm01.local ceph-mon[48597]: pgmap v1388: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:05.848 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:07:05.848 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:07:05.874 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:07:05.875 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:07:06.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:05 vm08.local ceph-mon[50506]: pgmap v1389: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:06.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:05 vm01.local ceph-mon[48597]: pgmap v1389: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:07 vm08.local ceph-mon[50506]: pgmap v1390: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:07:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:07 vm01.local ceph-mon[48597]: pgmap v1390: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:07:10.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:09 vm08.local ceph-mon[50506]: pgmap v1391: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:10.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:09 vm01.local ceph-mon[48597]: pgmap v1391: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:10.876 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:07:10.877 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:07:10.903 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:07:10.903 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:07:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:07:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:07:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:07:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:07:12.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:11 vm08.local ceph-mon[50506]: pgmap v1392: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:07:12.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:11 vm01.local ceph-mon[48597]: pgmap v1392: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:07:14.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:13 vm08.local ceph-mon[50506]: pgmap v1393: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:14.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:13 vm01.local ceph-mon[48597]: pgmap v1393: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:15.905 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:07:15.906 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:07:15.920 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:15 vm01.local ceph-mon[48597]: pgmap v1394: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:15.936 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:07:15.937 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:07:16.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:15 vm08.local ceph-mon[50506]: pgmap v1394: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:17 vm08.local ceph-mon[50506]: pgmap v1395: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:07:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:17 vm01.local ceph-mon[48597]: pgmap v1395: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:07:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:19 vm08.local ceph-mon[50506]: pgmap v1396: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:19 vm01.local ceph-mon[48597]: pgmap v1396: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:20.938 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:07:20.939 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:07:20.965 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:07:20.965 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:07:22.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:21 vm08.local ceph-mon[50506]: pgmap v1397: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:07:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:21 vm01.local ceph-mon[48597]: pgmap v1397: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:07:24.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:23 vm08.local ceph-mon[50506]: pgmap v1398: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:24.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:23 vm01.local ceph-mon[48597]: pgmap v1398: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:25.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:24 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:07:25.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:24 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:07:25.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:24 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:07:25.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:24 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:07:25.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:24 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:07:25.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:24 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:07:25.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:24 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:07:25.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:24 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:07:25.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:24 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:07:25.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:24 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:07:25.967 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:07:25.967 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:07:25.996 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:07:25.997 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:07:26.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:25 vm08.local ceph-mon[50506]: pgmap v1399: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:26.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:25 vm01.local ceph-mon[48597]: pgmap v1399: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:27 vm08.local ceph-mon[50506]: pgmap v1400: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:07:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:27 vm01.local ceph-mon[48597]: pgmap v1400: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:07:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:29 vm08.local ceph-mon[50506]: pgmap v1401: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:29 vm01.local ceph-mon[48597]: pgmap v1401: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:30.999 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:07:30.999 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:07:31.046 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:07:31.047 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:07:32.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:31 vm08.local ceph-mon[50506]: pgmap v1402: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:07:32.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:31 vm01.local ceph-mon[48597]: pgmap v1402: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:07:34.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:33 vm08.local ceph-mon[50506]: pgmap v1403: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:34.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:33 vm01.local ceph-mon[48597]: pgmap v1403: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:36.048 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:07:36.049 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:07:36.075 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:07:36.076 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:07:36.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:35 vm08.local ceph-mon[50506]: pgmap v1404: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:35 vm01.local ceph-mon[48597]: pgmap v1404: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:37 vm08.local ceph-mon[50506]: pgmap v1405: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:07:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:37 vm01.local ceph-mon[48597]: pgmap v1405: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:07:40.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:39 vm08.local ceph-mon[50506]: pgmap v1406: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:39 vm01.local ceph-mon[48597]: pgmap v1406: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:41.078 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:07:41.078 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:07:41.105 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:07:41.105 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:07:42.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:41 vm08.local ceph-mon[50506]: pgmap v1407: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:07:42.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:41 vm01.local ceph-mon[48597]: pgmap v1407: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:07:44.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:43 vm08.local ceph-mon[50506]: pgmap v1408: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:44.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:43 vm01.local ceph-mon[48597]: pgmap v1408: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:46.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:45 vm08.local ceph-mon[50506]: pgmap v1409: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:46.107 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:07:46.107 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:07:46.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:45 vm01.local ceph-mon[48597]: pgmap v1409: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:46.138 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:07:46.138 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:07:47.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:47 vm08.local ceph-mon[50506]: pgmap v1410: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:07:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:47 vm01.local ceph-mon[48597]: pgmap v1410: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:07:50.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:49 vm08.local ceph-mon[50506]: pgmap v1411: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:50.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:49 vm01.local ceph-mon[48597]: pgmap v1411: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:51.139 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:07:51.140 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:07:51.258 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:07:51.259 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:07:52.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:51 vm08.local ceph-mon[50506]: pgmap v1412: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:07:52.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:51 vm01.local ceph-mon[48597]: pgmap v1412: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:07:54.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:53 vm08.local ceph-mon[50506]: pgmap v1413: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:53 vm01.local ceph-mon[48597]: pgmap v1413: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:55.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:55 vm08.local ceph-mon[50506]: pgmap v1414: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:55 vm01.local ceph-mon[48597]: pgmap v1414: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:07:56.260 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:07:56.261 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:07:56.287 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:07:56.288 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:07:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:57 vm08.local ceph-mon[50506]: pgmap v1415: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:07:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:57 vm01.local ceph-mon[48597]: pgmap v1415: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:08:00.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:07:59 vm08.local ceph-mon[50506]: pgmap v1416: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:00.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:07:59 vm01.local ceph-mon[48597]: pgmap v1416: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:01.289 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:08:01.290 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:08:01.362 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:08:01.363 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:08:02.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:01 vm08.local ceph-mon[50506]: pgmap v1417: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:08:02.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:01 vm01.local ceph-mon[48597]: pgmap v1417: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:08:04.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:03 vm08.local ceph-mon[50506]: pgmap v1418: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:04.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:03 vm01.local ceph-mon[48597]: pgmap v1418: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:05.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:05 vm08.local ceph-mon[50506]: pgmap v1419: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:05.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:05 vm01.local ceph-mon[48597]: pgmap v1419: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:06.364 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:08:06.365 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:08:06.545 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:08:06.546 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:08:08.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:07 vm08.local ceph-mon[50506]: pgmap v1420: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:08:08.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:07 vm01.local ceph-mon[48597]: pgmap v1420: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:08:09.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:08 vm01.local ceph-mon[48597]: pgmap v1421: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:09.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:08 vm08.local ceph-mon[50506]: pgmap v1421: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:10.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:08:10.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:08:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:08:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:08:11.547 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:08:11.547 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:08:11.588 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:08:11.588 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:08:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:11 vm08.local ceph-mon[50506]: pgmap v1422: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:08:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:11 vm01.local ceph-mon[48597]: pgmap v1422: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:08:14.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:13 vm08.local ceph-mon[50506]: pgmap v1423: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:14.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:13 vm01.local ceph-mon[48597]: pgmap v1423: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:16.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:15 vm08.local ceph-mon[50506]: pgmap v1424: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:15 vm01.local ceph-mon[48597]: pgmap v1424: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:16.590 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:08:16.591 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:08:16.617 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:08:16.618 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:08:17.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:17 vm08.local ceph-mon[50506]: pgmap v1425: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:08:17.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:17 vm01.local ceph-mon[48597]: pgmap v1425: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:08:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:19 vm08.local ceph-mon[50506]: pgmap v1426: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:19 vm01.local ceph-mon[48597]: pgmap v1426: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:21.620 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:08:21.620 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:08:21.733 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:08:21.734 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:08:22.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:21 vm08.local ceph-mon[50506]: pgmap v1427: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:08:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:21 vm01.local ceph-mon[48597]: pgmap v1427: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:08:24.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:23 vm08.local ceph-mon[50506]: pgmap v1428: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:24.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:23 vm01.local ceph-mon[48597]: pgmap v1428: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:25.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:24 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:08:25.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:24 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:08:25.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:24 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:08:25.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:24 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:08:25.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:24 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:08:25.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:24 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:08:25.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:24 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:08:25.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:24 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:08:25.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:24 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:08:25.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:24 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:08:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:25 vm08.local ceph-mon[50506]: pgmap v1429: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:25 vm01.local ceph-mon[48597]: pgmap v1429: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:26.735 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:08:26.736 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:08:26.763 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:08:26.764 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:08:27.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:27 vm01.local ceph-mon[48597]: pgmap v1430: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:08:28.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:27 vm08.local ceph-mon[50506]: pgmap v1430: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:08:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:29 vm01.local ceph-mon[48597]: pgmap v1431: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:29 vm08.local ceph-mon[50506]: pgmap v1431: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:31.765 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:08:31.766 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:08:31.792 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:08:31.792 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:08:32.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:31 vm08.local ceph-mon[50506]: pgmap v1432: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:08:32.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:31 vm01.local ceph-mon[48597]: pgmap v1432: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:08:34.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:33 vm08.local ceph-mon[50506]: pgmap v1433: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:34.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:33 vm01.local ceph-mon[48597]: pgmap v1433: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:36.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:35 vm08.local ceph-mon[50506]: pgmap v1434: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:35 vm01.local ceph-mon[48597]: pgmap v1434: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:36.794 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:08:36.794 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:08:36.820 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:08:36.821 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:08:37.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:37 vm01.local ceph-mon[48597]: pgmap v1435: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:08:38.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:37 vm08.local ceph-mon[50506]: pgmap v1435: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:08:40.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:39 vm08.local ceph-mon[50506]: pgmap v1436: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:39 vm01.local ceph-mon[48597]: pgmap v1436: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:41.822 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:08:41.822 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:08:41.848 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:08:41.849 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:08:42.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:41 vm08.local ceph-mon[50506]: pgmap v1437: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:08:42.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:41 vm01.local ceph-mon[48597]: pgmap v1437: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:08:44.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:43 vm08.local ceph-mon[50506]: pgmap v1438: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:44.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:43 vm01.local ceph-mon[48597]: pgmap v1438: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:46.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:45 vm08.local ceph-mon[50506]: pgmap v1439: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:46.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:45 vm01.local ceph-mon[48597]: pgmap v1439: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:46.850 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:08:46.850 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:08:46.876 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:08:46.876 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:08:47.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:47 vm01.local ceph-mon[48597]: pgmap v1440: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:08:48.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:47 vm08.local ceph-mon[50506]: pgmap v1440: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:08:50.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:49 vm08.local ceph-mon[50506]: pgmap v1441: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:50.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:49 vm01.local ceph-mon[48597]: pgmap v1441: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:51.877 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:08:51.878 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:08:51.904 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:08:51.905 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:08:52.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:51 vm08.local ceph-mon[50506]: pgmap v1442: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:08:52.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:51 vm01.local ceph-mon[48597]: pgmap v1442: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:08:54.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:53 vm08.local ceph-mon[50506]: pgmap v1443: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:53 vm01.local ceph-mon[48597]: pgmap v1443: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:55 vm08.local ceph-mon[50506]: pgmap v1444: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:56.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:55 vm01.local ceph-mon[48597]: pgmap v1444: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:08:56.906 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:08:56.907 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:08:56.935 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:08:56.935 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:08:58.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:57 vm08.local ceph-mon[50506]: pgmap v1445: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:08:58.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:57 vm01.local ceph-mon[48597]: pgmap v1445: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:09:00.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:08:59 vm08.local ceph-mon[50506]: pgmap v1446: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:00.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:08:59 vm01.local ceph-mon[48597]: pgmap v1446: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:01.937 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:09:01.937 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:09:01.963 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:09:01.963 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:09:02.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:01 vm08.local ceph-mon[50506]: pgmap v1447: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:09:02.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:01 vm01.local ceph-mon[48597]: pgmap v1447: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:09:04.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:03 vm08.local ceph-mon[50506]: pgmap v1448: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:04.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:03 vm01.local ceph-mon[48597]: pgmap v1448: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:06.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:05 vm08.local ceph-mon[50506]: pgmap v1449: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:06.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:05 vm01.local ceph-mon[48597]: pgmap v1449: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:06.964 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:09:06.965 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:09:06.990 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:09:06.991 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:09:08.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:07 vm08.local ceph-mon[50506]: pgmap v1450: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:09:08.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:07 vm01.local ceph-mon[48597]: pgmap v1450: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:09:10.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:09 vm08.local ceph-mon[50506]: pgmap v1451: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:10.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:09 vm01.local ceph-mon[48597]: pgmap v1451: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:09:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:09:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:09:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:09:11.992 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:09:11.992 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:09:12.018 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:09:12.018 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:09:12.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:11 vm08.local ceph-mon[50506]: pgmap v1452: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:09:12.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:11 vm01.local ceph-mon[48597]: pgmap v1452: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:09:14.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:13 vm08.local ceph-mon[50506]: pgmap v1453: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:14.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:13 vm01.local ceph-mon[48597]: pgmap v1453: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:15.920 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:15 vm01.local ceph-mon[48597]: pgmap v1454: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:16.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:15 vm08.local ceph-mon[50506]: pgmap v1454: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:17.020 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:09:17.020 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:09:17.055 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:09:17.055 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:09:18.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:17 vm08.local ceph-mon[50506]: pgmap v1455: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:09:18.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:17 vm01.local ceph-mon[48597]: pgmap v1455: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:09:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:19 vm08.local ceph-mon[50506]: pgmap v1456: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:19 vm01.local ceph-mon[48597]: pgmap v1456: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:22.057 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:09:22.057 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:09:22.085 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:09:22.085 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:09:22.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:21 vm08.local ceph-mon[50506]: pgmap v1457: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:09:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:21 vm01.local ceph-mon[48597]: pgmap v1457: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:09:24.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:23 vm08.local ceph-mon[50506]: pgmap v1458: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:24.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:23 vm01.local ceph-mon[48597]: pgmap v1458: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:24.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:24 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:09:24.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:24 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:09:24.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:24 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:09:24.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:24 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:09:24.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:24 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:09:24.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:24 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:09:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:25 vm08.local ceph-mon[50506]: pgmap v1459: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:25 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:09:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:25 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:09:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:25 vm01.local ceph-mon[48597]: pgmap v1459: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:25 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:09:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:25 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:09:27.086 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:09:27.087 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:09:27.112 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:09:27.113 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:09:28.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:27 vm08.local ceph-mon[50506]: pgmap v1460: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:09:28.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:27 vm01.local ceph-mon[48597]: pgmap v1460: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:09:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:29 vm08.local ceph-mon[50506]: pgmap v1461: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:29 vm01.local ceph-mon[48597]: pgmap v1461: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:32.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:31 vm08.local ceph-mon[50506]: pgmap v1462: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:09:32.114 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:09:32.114 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:09:32.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:31 vm01.local ceph-mon[48597]: pgmap v1462: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:09:32.141 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:09:32.142 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:09:34.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:33 vm08.local ceph-mon[50506]: pgmap v1463: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:34.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:33 vm01.local ceph-mon[48597]: pgmap v1463: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:36.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:35 vm08.local ceph-mon[50506]: pgmap v1464: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:35 vm01.local ceph-mon[48597]: pgmap v1464: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:37.143 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:09:37.144 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:09:37.171 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:09:37.172 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:09:38.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:37 vm08.local ceph-mon[50506]: pgmap v1465: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:09:38.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:37 vm01.local ceph-mon[48597]: pgmap v1465: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:09:40.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:39 vm08.local ceph-mon[50506]: pgmap v1466: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:39 vm01.local ceph-mon[48597]: pgmap v1466: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:42.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:41 vm08.local ceph-mon[50506]: pgmap v1467: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:09:42.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:41 vm01.local ceph-mon[48597]: pgmap v1467: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:09:42.173 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:09:42.173 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:09:42.201 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:09:42.201 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:09:44.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:43 vm08.local ceph-mon[50506]: pgmap v1468: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:44.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:43 vm01.local ceph-mon[48597]: pgmap v1468: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:46.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:45 vm08.local ceph-mon[50506]: pgmap v1469: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:46.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:45 vm01.local ceph-mon[48597]: pgmap v1469: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:47.202 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:09:47.203 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:09:47.230 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:09:47.231 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:09:48.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:47 vm08.local ceph-mon[50506]: pgmap v1470: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:09:48.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:47 vm01.local ceph-mon[48597]: pgmap v1470: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:09:50.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:49 vm08.local ceph-mon[50506]: pgmap v1471: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:50.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:49 vm01.local ceph-mon[48597]: pgmap v1471: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:52.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:51 vm08.local ceph-mon[50506]: pgmap v1472: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:09:52.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:51 vm01.local ceph-mon[48597]: pgmap v1472: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:09:52.232 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:09:52.233 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:09:52.260 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:09:52.260 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:09:53.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:53 vm08.local ceph-mon[50506]: pgmap v1473: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:53.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:53 vm01.local ceph-mon[48597]: pgmap v1473: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:55 vm08.local ceph-mon[50506]: pgmap v1474: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:56.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:55 vm01.local ceph-mon[48597]: pgmap v1474: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:09:57.262 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:09:57.262 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:09:57.289 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:09:57.290 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:09:58.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:57 vm08.local ceph-mon[50506]: pgmap v1475: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:09:58.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:57 vm01.local ceph-mon[48597]: pgmap v1475: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:10:00.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:09:59 vm08.local ceph-mon[50506]: pgmap v1476: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:00.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:09:59 vm01.local ceph-mon[48597]: pgmap v1476: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:00.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:00 vm01.local ceph-mon[48597]: overall HEALTH_OK 2026-03-10T11:10:01.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:00 vm08.local ceph-mon[50506]: overall HEALTH_OK 2026-03-10T11:10:02.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:01 vm08.local ceph-mon[50506]: pgmap v1477: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:10:02.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:01 vm01.local ceph-mon[48597]: pgmap v1477: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:10:02.291 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:10:02.291 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:10:02.318 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:10:02.319 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:10:04.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:03 vm08.local ceph-mon[50506]: pgmap v1478: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:04.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:03 vm01.local ceph-mon[48597]: pgmap v1478: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:05.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:04 vm08.local ceph-mon[50506]: pgmap v1479: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:05.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:04 vm01.local ceph-mon[48597]: pgmap v1479: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:07.320 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:10:07.320 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:10:07.349 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:10:07.349 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:10:08.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:07 vm08.local ceph-mon[50506]: pgmap v1480: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:10:08.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:07 vm01.local ceph-mon[48597]: pgmap v1480: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:10:10.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:09 vm08.local ceph-mon[50506]: pgmap v1481: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:10.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:09 vm01.local ceph-mon[48597]: pgmap v1481: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:10:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:10:11.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:10:11.192 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:10:12.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:11 vm08.local ceph-mon[50506]: pgmap v1482: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:10:12.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:11 vm01.local ceph-mon[48597]: pgmap v1482: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:10:12.351 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:10:12.351 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:10:12.378 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:10:12.379 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:10:14.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:13 vm08.local ceph-mon[50506]: pgmap v1483: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:14.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:13 vm01.local ceph-mon[48597]: pgmap v1483: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:16.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:15 vm08.local ceph-mon[50506]: pgmap v1484: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:15 vm01.local ceph-mon[48597]: pgmap v1484: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:17.380 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:10:17.422 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:10:17.712 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:10:17.713 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:10:18.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:18 vm08.local ceph-mon[50506]: pgmap v1485: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:10:18.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:17 vm01.local ceph-mon[48597]: pgmap v1485: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:10:19.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:19 vm08.local ceph-mon[50506]: pgmap v1486: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:19 vm01.local ceph-mon[48597]: pgmap v1486: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:21 vm08.local ceph-mon[50506]: pgmap v1487: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:10:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:21 vm01.local ceph-mon[48597]: pgmap v1487: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:10:22.715 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:10:22.715 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:10:22.912 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:10:22.912 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:10:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:23 vm08.local ceph-mon[50506]: pgmap v1488: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:23.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:23 vm01.local ceph-mon[48597]: pgmap v1488: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:25 vm08.local ceph-mon[50506]: pgmap v1489: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:25 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:10:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:25 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:10:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:25 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:10:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:25 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:10:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:25 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:10:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:25 vm01.local ceph-mon[48597]: pgmap v1489: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:25 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:10:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:25 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:10:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:25 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:10:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:25 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:10:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:25 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:10:27.914 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:10:27.914 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:10:27.943 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:10:27.943 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:10:28.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:27 vm08.local ceph-mon[50506]: pgmap v1490: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:10:28.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:27 vm01.local ceph-mon[48597]: pgmap v1490: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:10:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:29 vm08.local ceph-mon[50506]: pgmap v1491: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:29 vm01.local ceph-mon[48597]: pgmap v1491: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:32.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:32 vm01.local ceph-mon[48597]: pgmap v1492: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:10:32.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:32 vm08.local ceph-mon[50506]: pgmap v1492: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:10:32.945 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:10:32.945 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:10:32.980 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:10:32.981 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:10:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:33 vm08.local ceph-mon[50506]: pgmap v1493: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:33.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:33 vm01.local ceph-mon[48597]: pgmap v1493: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:36.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:35 vm08.local ceph-mon[50506]: pgmap v1494: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:35 vm01.local ceph-mon[48597]: pgmap v1494: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:37.982 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:10:37.983 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:10:38.010 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:10:38.011 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:10:38.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:37 vm08.local ceph-mon[50506]: pgmap v1495: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:10:38.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:37 vm01.local ceph-mon[48597]: pgmap v1495: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:10:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:39 vm01.local ceph-mon[48597]: pgmap v1496: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:40.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:39 vm08.local ceph-mon[50506]: pgmap v1496: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:41.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:40 vm01.local ceph-mon[48597]: pgmap v1497: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:10:41.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:40 vm08.local ceph-mon[50506]: pgmap v1497: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:10:43.012 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:10:43.012 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:10:43.045 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:10:43.046 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:10:43.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:42 vm01.local ceph-mon[48597]: pgmap v1498: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:43.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:42 vm08.local ceph-mon[50506]: pgmap v1498: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:46.047 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:45 vm08.local ceph-mon[50506]: pgmap v1499: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:46.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:45 vm01.local ceph-mon[48597]: pgmap v1499: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:48.047 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:10:48.047 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:10:48.073 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:10:48.074 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:10:48.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:47 vm08.local ceph-mon[50506]: pgmap v1500: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:10:48.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:47 vm01.local ceph-mon[48597]: pgmap v1500: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:10:50.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:49 vm08.local ceph-mon[50506]: pgmap v1501: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:50.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:49 vm01.local ceph-mon[48597]: pgmap v1501: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:52.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:51 vm08.local ceph-mon[50506]: pgmap v1502: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:10:52.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:51 vm01.local ceph-mon[48597]: pgmap v1502: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:10:53.075 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:10:53.076 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:10:53.103 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:10:53.103 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:10:53.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:52 vm01.local ceph-mon[48597]: pgmap v1503: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:53.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:52 vm08.local ceph-mon[50506]: pgmap v1503: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:55 vm08.local ceph-mon[50506]: pgmap v1504: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:56.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:55 vm01.local ceph-mon[48597]: pgmap v1504: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:10:58.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:57 vm08.local ceph-mon[50506]: pgmap v1505: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:10:58.104 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:10:58.105 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:10:58.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:57 vm01.local ceph-mon[48597]: pgmap v1505: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:10:58.133 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:10:58.133 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:11:00.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:10:59 vm08.local ceph-mon[50506]: pgmap v1506: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:00.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:10:59 vm01.local ceph-mon[48597]: pgmap v1506: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:02.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:01 vm08.local ceph-mon[50506]: pgmap v1507: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:11:02.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:01 vm01.local ceph-mon[48597]: pgmap v1507: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:11:03.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:02 vm01.local ceph-mon[48597]: pgmap v1508: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:03.134 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:11:03.135 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:11:03.161 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:11:03.161 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:11:03.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:02 vm08.local ceph-mon[50506]: pgmap v1508: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:06.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:05 vm08.local ceph-mon[50506]: pgmap v1509: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:06.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:05 vm01.local ceph-mon[48597]: pgmap v1509: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:08.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:07 vm08.local ceph-mon[50506]: pgmap v1510: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:11:08.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:07 vm01.local ceph-mon[48597]: pgmap v1510: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:11:08.162 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:11:08.163 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:11:08.189 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:11:08.189 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:11:10.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:09 vm08.local ceph-mon[50506]: pgmap v1511: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:10.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:09 vm01.local ceph-mon[48597]: pgmap v1511: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:11:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:11:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:11:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:11:12.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:11 vm08.local ceph-mon[50506]: pgmap v1512: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:11:12.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:11 vm01.local ceph-mon[48597]: pgmap v1512: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:11:13.191 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:11:13.191 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:11:13.217 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:11:13.217 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:11:13.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:12 vm08.local ceph-mon[50506]: pgmap v1513: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:13.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:12 vm01.local ceph-mon[48597]: pgmap v1513: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:16.048 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:15 vm08.local ceph-mon[50506]: pgmap v1514: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:15 vm01.local ceph-mon[48597]: pgmap v1514: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:18.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:17 vm08.local ceph-mon[50506]: pgmap v1515: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:11:18.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:17 vm01.local ceph-mon[48597]: pgmap v1515: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:11:18.219 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:11:18.220 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:11:18.247 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:11:18.247 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:11:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:19 vm08.local ceph-mon[50506]: pgmap v1516: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:19 vm01.local ceph-mon[48597]: pgmap v1516: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:22.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:21 vm08.local ceph-mon[50506]: pgmap v1517: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:11:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:21 vm01.local ceph-mon[48597]: pgmap v1517: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:11:23.248 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:11:23.249 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:11:23.280 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:11:23.280 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:11:23.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:22 vm08.local ceph-mon[50506]: pgmap v1518: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:23.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:22 vm01.local ceph-mon[48597]: pgmap v1518: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:25 vm08.local ceph-mon[50506]: pgmap v1519: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:25 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:11:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:25 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:11:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:25 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:11:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:25 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:11:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:25 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:11:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:25 vm01.local ceph-mon[48597]: pgmap v1519: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:25 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:11:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:25 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:11:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:25 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:11:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:25 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:11:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:25 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:11:28.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:27 vm08.local ceph-mon[50506]: pgmap v1520: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:11:28.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:27 vm01.local ceph-mon[48597]: pgmap v1520: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:11:28.281 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:11:28.282 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:11:28.308 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:11:28.308 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:11:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:29 vm08.local ceph-mon[50506]: pgmap v1521: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:29 vm01.local ceph-mon[48597]: pgmap v1521: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:32.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:31 vm08.local ceph-mon[50506]: pgmap v1522: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:11:32.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:31 vm01.local ceph-mon[48597]: pgmap v1522: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:11:33.309 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:11:33.310 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:11:33.335 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:11:33.336 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:11:33.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:32 vm08.local ceph-mon[50506]: pgmap v1523: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:33.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:32 vm01.local ceph-mon[48597]: pgmap v1523: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:36.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:35 vm08.local ceph-mon[50506]: pgmap v1524: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:35 vm01.local ceph-mon[48597]: pgmap v1524: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:38.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:37 vm08.local ceph-mon[50506]: pgmap v1525: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:11:38.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:37 vm01.local ceph-mon[48597]: pgmap v1525: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:11:38.337 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:11:38.337 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:11:38.462 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:11:38.463 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:11:40.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:39 vm08.local ceph-mon[50506]: pgmap v1526: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:39 vm01.local ceph-mon[48597]: pgmap v1526: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:42.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:41 vm08.local ceph-mon[50506]: pgmap v1527: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:11:42.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:41 vm01.local ceph-mon[48597]: pgmap v1527: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:11:43.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:42 vm08.local ceph-mon[50506]: pgmap v1528: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:43.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:42 vm01.local ceph-mon[48597]: pgmap v1528: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:43.464 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:11:43.465 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:11:43.491 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:11:43.491 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:11:46.048 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:45 vm08.local ceph-mon[50506]: pgmap v1529: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:46.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:45 vm01.local ceph-mon[48597]: pgmap v1529: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:47 vm08.local ceph-mon[50506]: pgmap v1530: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:11:48.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:47 vm01.local ceph-mon[48597]: pgmap v1530: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:11:48.492 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:11:48.493 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:11:48.520 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:11:48.521 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:11:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:48 vm08.local ceph-mon[50506]: pgmap v1531: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:49.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:48 vm01.local ceph-mon[48597]: pgmap v1531: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:52.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:51 vm08.local ceph-mon[50506]: pgmap v1532: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:11:52.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:51 vm01.local ceph-mon[48597]: pgmap v1532: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:11:53.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:52 vm08.local ceph-mon[50506]: pgmap v1533: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:53.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:52 vm01.local ceph-mon[48597]: pgmap v1533: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:53.522 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:11:53.523 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:11:53.549 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:11:53.550 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:11:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:55 vm08.local ceph-mon[50506]: pgmap v1534: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:56.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:55 vm01.local ceph-mon[48597]: pgmap v1534: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:11:58.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:57 vm08.local ceph-mon[50506]: pgmap v1535: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:11:58.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:57 vm01.local ceph-mon[48597]: pgmap v1535: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:11:58.551 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:11:58.551 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:11:58.579 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:11:58.579 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:12:00.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:11:59 vm08.local ceph-mon[50506]: pgmap v1536: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:00.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:11:59 vm01.local ceph-mon[48597]: pgmap v1536: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:02.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:01 vm08.local ceph-mon[50506]: pgmap v1537: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:12:02.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:01 vm01.local ceph-mon[48597]: pgmap v1537: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:12:03.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:02 vm08.local ceph-mon[50506]: pgmap v1538: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:03.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:02 vm01.local ceph-mon[48597]: pgmap v1538: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:03.581 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:12:03.581 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:12:03.608 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:12:03.608 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:12:06.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:05 vm08.local ceph-mon[50506]: pgmap v1539: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:06.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:05 vm01.local ceph-mon[48597]: pgmap v1539: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:08.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:07 vm08.local ceph-mon[50506]: pgmap v1540: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:12:08.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:07 vm01.local ceph-mon[48597]: pgmap v1540: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:12:08.610 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:12:08.610 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:12:08.638 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:12:08.639 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:12:10.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:09 vm08.local ceph-mon[50506]: pgmap v1541: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:10.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:09 vm01.local ceph-mon[48597]: pgmap v1541: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:12:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:12:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:12:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:12:12.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:11 vm08.local ceph-mon[50506]: pgmap v1542: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:12:12.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:11 vm01.local ceph-mon[48597]: pgmap v1542: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:12:13.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:12 vm08.local ceph-mon[50506]: pgmap v1543: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:13.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:12 vm01.local ceph-mon[48597]: pgmap v1543: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:13.640 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:12:13.641 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:12:13.667 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:12:13.668 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:12:16.047 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:15 vm08.local ceph-mon[50506]: pgmap v1544: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:15 vm01.local ceph-mon[48597]: pgmap v1544: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:18.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:17 vm08.local ceph-mon[50506]: pgmap v1545: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:12:18.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:17 vm01.local ceph-mon[48597]: pgmap v1545: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:12:18.669 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:12:18.670 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:12:18.697 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:12:18.698 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:12:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:19 vm08.local ceph-mon[50506]: pgmap v1546: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:19 vm01.local ceph-mon[48597]: pgmap v1546: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:22.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:21 vm08.local ceph-mon[50506]: pgmap v1547: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:12:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:21 vm01.local ceph-mon[48597]: pgmap v1547: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:12:23.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:22 vm08.local ceph-mon[50506]: pgmap v1548: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:23.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:22 vm01.local ceph-mon[48597]: pgmap v1548: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:23.700 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:12:23.700 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:12:23.726 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:12:23.726 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:12:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:25 vm01.local ceph-mon[48597]: pgmap v1549: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:25 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:12:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:25 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:12:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:25 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:12:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:25 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:12:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:25 vm08.local ceph-mon[50506]: pgmap v1549: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:25 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:12:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:25 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:12:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:25 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:12:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:25 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:12:27.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:26 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:12:27.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:26 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:12:27.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:26 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:12:27.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:26 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:12:28.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:27 vm01.local ceph-mon[48597]: pgmap v1550: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:12:28.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:27 vm08.local ceph-mon[50506]: pgmap v1550: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:12:28.728 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:12:28.728 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:12:28.755 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:12:28.755 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:12:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:29 vm01.local ceph-mon[48597]: pgmap v1551: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:30.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:29 vm08.local ceph-mon[50506]: pgmap v1551: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:32.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:31 vm01.local ceph-mon[48597]: pgmap v1552: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:12:32.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:31 vm08.local ceph-mon[50506]: pgmap v1552: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:12:33.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:32 vm08.local ceph-mon[50506]: pgmap v1553: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:33.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:32 vm01.local ceph-mon[48597]: pgmap v1553: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:33.757 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:12:33.757 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:12:33.784 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:12:33.785 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:12:36.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:35 vm08.local ceph-mon[50506]: pgmap v1554: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:35 vm01.local ceph-mon[48597]: pgmap v1554: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:38.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:37 vm08.local ceph-mon[50506]: pgmap v1555: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:12:38.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:37 vm01.local ceph-mon[48597]: pgmap v1555: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:12:38.787 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:12:38.787 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:12:38.813 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:12:38.814 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:12:40.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:39 vm08.local ceph-mon[50506]: pgmap v1556: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:39 vm01.local ceph-mon[48597]: pgmap v1556: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:41.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:40 vm08.local ceph-mon[50506]: pgmap v1557: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:12:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:40 vm01.local ceph-mon[48597]: pgmap v1557: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:12:43.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:43 vm08.local ceph-mon[50506]: pgmap v1558: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:43.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:43 vm01.local ceph-mon[48597]: pgmap v1558: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:43.816 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:12:43.816 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:12:43.842 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:12:43.843 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:12:46.048 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:45 vm08.local ceph-mon[50506]: pgmap v1559: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:46.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:45 vm01.local ceph-mon[48597]: pgmap v1559: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:48.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:47 vm08.local ceph-mon[50506]: pgmap v1560: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:12:48.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:47 vm01.local ceph-mon[48597]: pgmap v1560: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:12:48.844 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:12:48.844 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:12:48.872 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:12:48.872 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:12:50.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:49 vm08.local ceph-mon[50506]: pgmap v1561: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:50.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:49 vm01.local ceph-mon[48597]: pgmap v1561: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:52.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:51 vm08.local ceph-mon[50506]: pgmap v1562: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:12:52.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:51 vm01.local ceph-mon[48597]: pgmap v1562: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:12:53.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:53 vm08.local ceph-mon[50506]: pgmap v1563: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:53.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:53 vm01.local ceph-mon[48597]: pgmap v1563: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:53.874 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:12:53.874 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:12:53.900 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:12:53.901 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:12:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:55 vm08.local ceph-mon[50506]: pgmap v1564: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:56.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:55 vm01.local ceph-mon[48597]: pgmap v1564: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:12:58.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:57 vm08.local ceph-mon[50506]: pgmap v1565: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:12:58.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:57 vm01.local ceph-mon[48597]: pgmap v1565: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:12:58.902 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:12:58.903 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:12:58.928 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:12:58.928 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:13:00.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:12:59 vm08.local ceph-mon[50506]: pgmap v1566: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:00.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:12:59 vm01.local ceph-mon[48597]: pgmap v1566: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:02.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:01 vm08.local ceph-mon[50506]: pgmap v1567: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:13:02.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:01 vm01.local ceph-mon[48597]: pgmap v1567: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:13:03.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:03 vm08.local ceph-mon[50506]: pgmap v1568: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:03.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:03 vm01.local ceph-mon[48597]: pgmap v1568: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:03.930 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:13:03.930 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:13:03.958 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:13:03.959 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:13:06.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:05 vm08.local ceph-mon[50506]: pgmap v1569: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:06.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:05 vm01.local ceph-mon[48597]: pgmap v1569: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:08.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:07 vm08.local ceph-mon[50506]: pgmap v1570: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:13:08.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:07 vm01.local ceph-mon[48597]: pgmap v1570: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:13:08.961 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:13:08.961 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:13:08.987 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:13:08.988 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:13:10.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:09 vm01.local ceph-mon[48597]: pgmap v1571: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:10.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:09 vm08.local ceph-mon[50506]: pgmap v1571: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:11.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:13:11.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:13:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:13:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:13:12.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:11 vm01.local ceph-mon[48597]: pgmap v1572: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:13:12.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:11 vm08.local ceph-mon[50506]: pgmap v1572: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:13:13.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:13 vm08.local ceph-mon[50506]: pgmap v1573: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:13.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:13 vm01.local ceph-mon[48597]: pgmap v1573: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:13.989 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:13:13.990 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:13:14.025 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:13:14.025 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:13:16.047 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:15 vm08.local ceph-mon[50506]: pgmap v1574: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:15 vm01.local ceph-mon[48597]: pgmap v1574: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:18.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:17 vm08.local ceph-mon[50506]: pgmap v1575: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:13:18.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:17 vm01.local ceph-mon[48597]: pgmap v1575: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:13:19.027 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:13:19.027 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:13:19.055 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:13:19.055 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:13:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:19 vm08.local ceph-mon[50506]: pgmap v1576: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:19 vm01.local ceph-mon[48597]: pgmap v1576: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:22.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:21 vm08.local ceph-mon[50506]: pgmap v1577: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:13:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:21 vm01.local ceph-mon[48597]: pgmap v1577: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:13:23.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:23 vm08.local ceph-mon[50506]: pgmap v1578: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:23.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:23 vm01.local ceph-mon[48597]: pgmap v1578: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:24.057 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:13:24.057 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:13:24.092 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:13:24.093 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:13:26.008 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:25 vm08.local ceph-mon[50506]: pgmap v1579: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:26.070 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:25 vm01.local ceph-mon[48597]: pgmap v1579: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:27.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:26 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:13:27.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:26 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:13:27.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:26 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:13:27.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:26 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:13:27.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:26 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:13:27.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:26 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:13:27.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:26 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:13:27.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:26 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:13:27.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:26 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:13:27.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:26 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:13:27.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:26 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:13:27.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:26 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:13:27.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:26 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:13:27.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:26 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:13:28.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:27 vm08.local ceph-mon[50506]: pgmap v1580: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:13:28.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:27 vm01.local ceph-mon[48597]: pgmap v1580: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:13:29.094 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:13:29.095 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:13:29.120 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:13:29.121 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:13:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:29 vm08.local ceph-mon[50506]: pgmap v1581: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:29 vm01.local ceph-mon[48597]: pgmap v1581: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:32.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:31 vm08.local ceph-mon[50506]: pgmap v1582: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:13:32.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:31 vm01.local ceph-mon[48597]: pgmap v1582: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:13:33.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:33 vm08.local ceph-mon[50506]: pgmap v1583: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:33.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:33 vm01.local ceph-mon[48597]: pgmap v1583: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:34.123 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:13:34.123 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:13:34.150 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:13:34.151 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:13:36.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:35 vm08.local ceph-mon[50506]: pgmap v1584: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:35 vm01.local ceph-mon[48597]: pgmap v1584: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:38.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:37 vm08.local ceph-mon[50506]: pgmap v1585: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:13:38.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:37 vm01.local ceph-mon[48597]: pgmap v1585: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:13:39.153 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:13:39.154 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:13:39.200 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:13:39.200 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:13:40.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:39 vm08.local ceph-mon[50506]: pgmap v1586: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:39 vm01.local ceph-mon[48597]: pgmap v1586: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:42.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:41 vm08.local ceph-mon[50506]: pgmap v1587: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:13:42.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:41 vm01.local ceph-mon[48597]: pgmap v1587: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:13:43.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:43 vm08.local ceph-mon[50506]: pgmap v1588: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:43.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:43 vm01.local ceph-mon[48597]: pgmap v1588: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:44.201 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:13:44.202 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:13:44.228 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:13:44.228 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:13:46.047 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:45 vm08.local ceph-mon[50506]: pgmap v1589: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:46.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:45 vm01.local ceph-mon[48597]: pgmap v1589: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:48.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:47 vm08.local ceph-mon[50506]: pgmap v1590: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:13:48.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:47 vm01.local ceph-mon[48597]: pgmap v1590: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:13:49.230 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:13:49.230 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:13:49.259 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:13:49.260 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:13:50.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:49 vm08.local ceph-mon[50506]: pgmap v1591: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:50.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:49 vm01.local ceph-mon[48597]: pgmap v1591: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:52.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:51 vm08.local ceph-mon[50506]: pgmap v1592: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:13:52.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:51 vm01.local ceph-mon[48597]: pgmap v1592: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:13:53.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:53 vm08.local ceph-mon[50506]: pgmap v1593: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:53.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:53 vm01.local ceph-mon[48597]: pgmap v1593: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:54.262 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:13:54.262 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:13:54.290 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:13:54.291 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:13:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:55 vm08.local ceph-mon[50506]: pgmap v1594: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:56.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:55 vm01.local ceph-mon[48597]: pgmap v1594: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:13:57.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:57 vm01.local ceph-mon[48597]: pgmap v1595: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:13:58.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:57 vm08.local ceph-mon[50506]: pgmap v1595: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:13:59.292 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:13:59.293 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:13:59.322 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:13:59.322 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:14:00.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:13:59 vm08.local ceph-mon[50506]: pgmap v1596: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:00.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:13:59 vm01.local ceph-mon[48597]: pgmap v1596: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:02.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:01 vm08.local ceph-mon[50506]: pgmap v1597: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:14:02.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:01 vm01.local ceph-mon[48597]: pgmap v1597: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:14:03.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:03 vm08.local ceph-mon[50506]: pgmap v1598: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:03.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:03 vm01.local ceph-mon[48597]: pgmap v1598: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:04.324 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:14:04.324 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:14:04.349 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:14:04.350 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:14:06.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:05 vm08.local ceph-mon[50506]: pgmap v1599: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:06.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:05 vm01.local ceph-mon[48597]: pgmap v1599: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:08.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:07 vm08.local ceph-mon[50506]: pgmap v1600: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:14:08.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:07 vm01.local ceph-mon[48597]: pgmap v1600: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:14:09.351 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:14:09.351 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:14:09.376 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:14:09.377 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:14:10.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:09 vm08.local ceph-mon[50506]: pgmap v1601: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:10.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:09 vm01.local ceph-mon[48597]: pgmap v1601: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:14:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:14:11.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:14:11.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:14:12.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:11 vm08.local ceph-mon[50506]: pgmap v1602: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:14:12.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:11 vm01.local ceph-mon[48597]: pgmap v1602: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:14:13.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:13 vm08.local ceph-mon[50506]: pgmap v1603: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:13.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:13 vm01.local ceph-mon[48597]: pgmap v1603: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:14.378 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:14:14.378 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:14:14.405 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:14:14.405 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:14:16.047 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:15 vm08.local ceph-mon[50506]: pgmap v1604: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:15 vm01.local ceph-mon[48597]: pgmap v1604: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:18.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:17 vm08.local ceph-mon[50506]: pgmap v1605: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:14:18.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:17 vm01.local ceph-mon[48597]: pgmap v1605: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:14:19.407 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:14:19.407 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:14:19.435 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:14:19.436 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:14:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:19 vm08.local ceph-mon[50506]: pgmap v1606: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:19 vm01.local ceph-mon[48597]: pgmap v1606: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:22.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:21 vm08.local ceph-mon[50506]: pgmap v1607: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:14:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:21 vm01.local ceph-mon[48597]: pgmap v1607: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:14:23.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:23 vm08.local ceph-mon[50506]: pgmap v1608: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:23.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:23 vm01.local ceph-mon[48597]: pgmap v1608: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:24.437 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:14:24.438 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:14:24.464 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:14:24.465 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:14:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:25 vm08.local ceph-mon[50506]: pgmap v1609: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:25 vm01.local ceph-mon[48597]: pgmap v1609: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:26.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:26 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:14:26.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:26 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:14:26.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:26 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:14:26.907 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:26 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:14:26.907 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:26 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:14:26.907 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:26 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:14:27.830 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:27 vm08.local ceph-mon[50506]: pgmap v1610: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:14:27.830 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:27 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:14:27.830 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:27 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:14:27.830 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:27 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:14:27.977 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:27 vm01.local ceph-mon[48597]: pgmap v1610: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:14:27.977 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:27 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:14:27.977 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:27 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:14:27.977 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:27 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:14:29.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:28 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:14:29.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:28 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:14:29.345 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:28 vm08.local ceph-mon[50506]: pgmap v1611: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:29.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:28 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:14:29.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:28 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:14:29.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:28 vm01.local ceph-mon[48597]: pgmap v1611: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:29.466 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:14:29.467 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:14:29.494 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:14:29.495 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:14:32.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:31 vm08.local ceph-mon[50506]: pgmap v1612: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:14:32.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:31 vm01.local ceph-mon[48597]: pgmap v1612: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:14:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:33 vm08.local ceph-mon[50506]: pgmap v1613: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:33.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:33 vm01.local ceph-mon[48597]: pgmap v1613: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:34.497 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:14:34.497 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:14:34.525 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:14:34.525 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:14:36.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:35 vm08.local ceph-mon[50506]: pgmap v1614: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:35 vm01.local ceph-mon[48597]: pgmap v1614: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:38.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:37 vm08.local ceph-mon[50506]: pgmap v1615: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:14:38.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:37 vm01.local ceph-mon[48597]: pgmap v1615: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:14:39.526 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:14:39.527 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:14:39.553 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:14:39.553 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:14:40.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:39 vm08.local ceph-mon[50506]: pgmap v1616: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:39 vm01.local ceph-mon[48597]: pgmap v1616: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:42.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:41 vm08.local ceph-mon[50506]: pgmap v1617: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:14:42.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:41 vm01.local ceph-mon[48597]: pgmap v1617: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:14:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:43 vm08.local ceph-mon[50506]: pgmap v1618: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:43.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:43 vm01.local ceph-mon[48597]: pgmap v1618: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:44.555 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:14:44.555 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:14:44.580 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:14:44.581 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:14:46.062 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:45 vm01.local ceph-mon[48597]: pgmap v1619: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:46.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:45 vm08.local ceph-mon[50506]: pgmap v1619: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:48.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:47 vm08.local ceph-mon[50506]: pgmap v1620: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:14:48.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:47 vm01.local ceph-mon[48597]: pgmap v1620: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:14:49.583 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:14:49.584 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:14:49.612 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:14:49.613 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:14:50.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:49 vm08.local ceph-mon[50506]: pgmap v1621: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:50.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:49 vm01.local ceph-mon[48597]: pgmap v1621: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:52.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:51 vm08.local ceph-mon[50506]: pgmap v1622: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:14:52.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:51 vm01.local ceph-mon[48597]: pgmap v1622: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:14:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:53 vm08.local ceph-mon[50506]: pgmap v1623: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:53.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:53 vm01.local ceph-mon[48597]: pgmap v1623: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:54.614 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:14:54.615 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:14:54.643 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:14:54.643 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:14:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:55 vm08.local ceph-mon[50506]: pgmap v1624: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:56.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:55 vm01.local ceph-mon[48597]: pgmap v1624: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:14:58.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:57 vm08.local ceph-mon[50506]: pgmap v1625: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:14:58.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:57 vm01.local ceph-mon[48597]: pgmap v1625: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:14:59.645 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:14:59.645 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:14:59.673 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:14:59.674 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:15:00.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:14:59 vm08.local ceph-mon[50506]: pgmap v1626: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:00.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:14:59 vm01.local ceph-mon[48597]: pgmap v1626: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:02.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:01 vm08.local ceph-mon[50506]: pgmap v1627: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:15:02.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:01 vm01.local ceph-mon[48597]: pgmap v1627: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:15:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:03 vm08.local ceph-mon[50506]: pgmap v1628: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:03.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:03 vm01.local ceph-mon[48597]: pgmap v1628: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:04.675 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:15:04.676 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:15:04.703 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:15:04.704 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:15:06.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:05 vm08.local ceph-mon[50506]: pgmap v1629: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:06.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:05 vm01.local ceph-mon[48597]: pgmap v1629: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:08.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:07 vm08.local ceph-mon[50506]: pgmap v1630: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:15:08.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:07 vm01.local ceph-mon[48597]: pgmap v1630: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:15:09.705 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:15:09.705 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:15:09.731 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:15:09.732 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:15:10.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:09 vm08.local ceph-mon[50506]: pgmap v1631: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:10.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:09 vm01.local ceph-mon[48597]: pgmap v1631: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:15:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:15:11.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:15:11.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:15:12.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:11 vm08.local ceph-mon[50506]: pgmap v1632: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:15:12.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:11 vm01.local ceph-mon[48597]: pgmap v1632: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:15:13.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:13 vm08.local ceph-mon[50506]: pgmap v1633: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:13.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:13 vm01.local ceph-mon[48597]: pgmap v1633: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:14.733 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:15:14.734 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:15:14.760 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:15:14.760 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:15:16.059 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:15 vm01.local ceph-mon[48597]: pgmap v1634: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:16.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:15 vm08.local ceph-mon[50506]: pgmap v1634: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:18.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:17 vm08.local ceph-mon[50506]: pgmap v1635: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:15:18.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:17 vm01.local ceph-mon[48597]: pgmap v1635: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:15:19.762 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:15:19.762 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:15:19.789 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:15:19.789 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:15:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:19 vm08.local ceph-mon[50506]: pgmap v1636: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:19 vm01.local ceph-mon[48597]: pgmap v1636: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:22.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:21 vm08.local ceph-mon[50506]: pgmap v1637: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:15:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:21 vm01.local ceph-mon[48597]: pgmap v1637: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:15:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:23 vm08.local ceph-mon[50506]: pgmap v1638: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:23.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:23 vm01.local ceph-mon[48597]: pgmap v1638: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:24.791 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:15:24.791 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:15:24.824 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:15:24.825 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:15:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:25 vm01.local ceph-mon[48597]: pgmap v1639: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:25 vm08.local ceph-mon[50506]: pgmap v1639: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:28.019 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:27 vm01.local ceph-mon[48597]: pgmap v1640: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:15:28.021 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:27 vm08.local ceph-mon[50506]: pgmap v1640: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:15:29.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:28 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:15:29.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:28 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:15:29.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:28 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:15:29.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:28 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:15:29.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:28 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:15:29.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:28 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:15:29.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:28 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:15:29.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:28 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:15:29.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:28 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:15:29.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:28 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:15:29.826 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:15:29.827 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:15:29.854 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:15:29.854 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:15:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:29 vm08.local ceph-mon[50506]: pgmap v1641: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:29 vm01.local ceph-mon[48597]: pgmap v1641: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:32.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:31 vm08.local ceph-mon[50506]: pgmap v1642: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:15:32.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:31 vm01.local ceph-mon[48597]: pgmap v1642: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:15:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:33 vm08.local ceph-mon[50506]: pgmap v1643: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:33.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:33 vm01.local ceph-mon[48597]: pgmap v1643: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:34.856 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:15:34.857 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:15:34.883 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:15:34.883 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:15:36.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:35 vm08.local ceph-mon[50506]: pgmap v1644: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:35 vm01.local ceph-mon[48597]: pgmap v1644: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:38.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:37 vm08.local ceph-mon[50506]: pgmap v1645: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:15:38.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:37 vm01.local ceph-mon[48597]: pgmap v1645: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:15:39.885 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:15:39.886 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:15:39.968 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:15:39.969 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:15:40.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:39 vm08.local ceph-mon[50506]: pgmap v1646: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:39 vm01.local ceph-mon[48597]: pgmap v1646: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:42.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:41 vm08.local ceph-mon[50506]: pgmap v1647: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:15:42.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:41 vm01.local ceph-mon[48597]: pgmap v1647: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:15:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:43 vm08.local ceph-mon[50506]: pgmap v1648: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:43.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:43 vm01.local ceph-mon[48597]: pgmap v1648: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:44.970 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:15:44.970 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:15:45.045 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:15:45.046 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:15:46.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:45 vm08.local ceph-mon[50506]: pgmap v1649: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:46.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:45 vm01.local ceph-mon[48597]: pgmap v1649: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:48.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:47 vm08.local ceph-mon[50506]: pgmap v1650: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:15:48.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:47 vm01.local ceph-mon[48597]: pgmap v1650: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:15:50.047 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:15:50.048 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:15:50.075 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:15:50.076 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:15:50.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:49 vm08.local ceph-mon[50506]: pgmap v1651: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:50.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:49 vm01.local ceph-mon[48597]: pgmap v1651: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:52.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:51 vm08.local ceph-mon[50506]: pgmap v1652: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:15:52.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:51 vm01.local ceph-mon[48597]: pgmap v1652: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:15:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:53 vm08.local ceph-mon[50506]: pgmap v1653: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:53.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:53 vm01.local ceph-mon[48597]: pgmap v1653: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:55.078 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:15:55.078 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:15:55.105 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:15:55.106 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:15:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:55 vm08.local ceph-mon[50506]: pgmap v1654: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:56.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:55 vm01.local ceph-mon[48597]: pgmap v1654: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:15:58.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:57 vm08.local ceph-mon[50506]: pgmap v1655: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:15:58.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:57 vm01.local ceph-mon[48597]: pgmap v1655: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:16:00.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:15:59 vm08.local ceph-mon[50506]: pgmap v1656: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:00.107 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:16:00.108 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:16:00.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:15:59 vm01.local ceph-mon[48597]: pgmap v1656: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:00.134 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:16:00.134 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:16:02.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:01 vm08.local ceph-mon[50506]: pgmap v1657: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:16:02.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:01 vm01.local ceph-mon[48597]: pgmap v1657: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:16:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:03 vm08.local ceph-mon[50506]: pgmap v1658: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:03.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:03 vm01.local ceph-mon[48597]: pgmap v1658: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:05.136 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:16:05.136 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:16:05.162 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:16:05.163 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:16:06.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:05 vm08.local ceph-mon[50506]: pgmap v1659: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:06.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:05 vm01.local ceph-mon[48597]: pgmap v1659: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:08.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:07 vm08.local ceph-mon[50506]: pgmap v1660: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:16:08.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:07 vm01.local ceph-mon[48597]: pgmap v1660: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:16:10.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:09 vm08.local ceph-mon[50506]: pgmap v1661: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:10.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:09 vm01.local ceph-mon[48597]: pgmap v1661: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:10.167 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:16:10.167 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:16:10.193 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:16:10.194 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:16:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:16:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:16:11.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:16:11.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:16:12.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:11 vm08.local ceph-mon[50506]: pgmap v1662: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:16:12.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:11 vm01.local ceph-mon[48597]: pgmap v1662: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:16:13.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:13 vm08.local ceph-mon[50506]: pgmap v1663: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:13.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:13 vm01.local ceph-mon[48597]: pgmap v1663: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:15.195 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:16:15.196 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:16:15.225 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:16:15.225 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:16:16.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:15 vm08.local ceph-mon[50506]: pgmap v1664: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:15 vm01.local ceph-mon[48597]: pgmap v1664: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:18.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:17 vm08.local ceph-mon[50506]: pgmap v1665: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:16:18.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:17 vm01.local ceph-mon[48597]: pgmap v1665: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:16:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:19 vm08.local ceph-mon[50506]: pgmap v1666: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:19 vm01.local ceph-mon[48597]: pgmap v1666: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:20.227 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:16:20.227 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:16:20.263 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:16:20.264 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:16:22.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:21 vm08.local ceph-mon[50506]: pgmap v1667: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:16:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:21 vm01.local ceph-mon[48597]: pgmap v1667: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:16:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:23 vm08.local ceph-mon[50506]: pgmap v1668: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:23.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:23 vm01.local ceph-mon[48597]: pgmap v1668: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:25.265 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:16:25.266 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:16:25.294 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:16:25.294 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:16:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:25 vm08.local ceph-mon[50506]: pgmap v1669: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:25 vm01.local ceph-mon[48597]: pgmap v1669: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:28.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:27 vm08.local ceph-mon[50506]: pgmap v1670: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:16:28.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:27 vm01.local ceph-mon[48597]: pgmap v1670: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:16:29.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:28 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:16:29.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:28 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:16:29.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:28 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:16:29.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:28 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:16:29.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:28 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:16:29.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:28 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:16:29.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:28 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:16:29.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:28 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:16:29.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:28 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:16:29.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:28 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:16:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:29 vm08.local ceph-mon[50506]: pgmap v1671: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:29 vm01.local ceph-mon[48597]: pgmap v1671: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:30.296 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:16:30.296 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:16:30.322 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:16:30.322 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:16:32.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:31 vm08.local ceph-mon[50506]: pgmap v1672: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:16:32.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:31 vm01.local ceph-mon[48597]: pgmap v1672: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:16:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:33 vm08.local ceph-mon[50506]: pgmap v1673: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:33.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:33 vm01.local ceph-mon[48597]: pgmap v1673: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:35.324 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:16:35.324 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:16:35.357 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:16:35.358 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:16:36.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:35 vm08.local ceph-mon[50506]: pgmap v1674: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:35 vm01.local ceph-mon[48597]: pgmap v1674: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:38.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:37 vm08.local ceph-mon[50506]: pgmap v1675: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:16:38.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:37 vm01.local ceph-mon[48597]: pgmap v1675: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:16:40.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:39 vm08.local ceph-mon[50506]: pgmap v1676: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:39 vm01.local ceph-mon[48597]: pgmap v1676: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:40.360 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:16:40.360 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:16:40.388 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:16:40.389 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:16:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:41 vm08.local ceph-mon[50506]: pgmap v1677: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:16:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:41 vm01.local ceph-mon[48597]: pgmap v1677: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:16:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:43 vm08.local ceph-mon[50506]: pgmap v1678: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:43.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:43 vm01.local ceph-mon[48597]: pgmap v1678: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:45.391 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:16:45.391 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:16:45.419 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:16:45.420 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:16:46.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:45 vm08.local ceph-mon[50506]: pgmap v1679: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:46.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:45 vm01.local ceph-mon[48597]: pgmap v1679: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:48.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:47 vm08.local ceph-mon[50506]: pgmap v1680: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:16:48.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:47 vm01.local ceph-mon[48597]: pgmap v1680: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:16:50.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:49 vm08.local ceph-mon[50506]: pgmap v1681: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:50.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:49 vm01.local ceph-mon[48597]: pgmap v1681: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:50.422 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:16:50.422 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:16:50.448 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:16:50.449 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:16:52.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:51 vm08.local ceph-mon[50506]: pgmap v1682: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:16:52.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:51 vm01.local ceph-mon[48597]: pgmap v1682: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:16:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:53 vm08.local ceph-mon[50506]: pgmap v1683: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:53.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:53 vm01.local ceph-mon[48597]: pgmap v1683: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:55.450 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:16:55.451 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:16:55.475 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:16:55.476 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:16:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:55 vm08.local ceph-mon[50506]: pgmap v1684: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:56.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:55 vm01.local ceph-mon[48597]: pgmap v1684: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:16:58.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:57 vm08.local ceph-mon[50506]: pgmap v1685: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:16:58.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:57 vm01.local ceph-mon[48597]: pgmap v1685: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:17:00.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:16:59 vm08.local ceph-mon[50506]: pgmap v1686: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:00.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:16:59 vm01.local ceph-mon[48597]: pgmap v1686: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:00.477 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:17:00.478 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:17:00.503 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:17:00.503 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:17:02.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:01 vm08.local ceph-mon[50506]: pgmap v1687: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:17:02.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:01 vm01.local ceph-mon[48597]: pgmap v1687: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:17:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:03 vm08.local ceph-mon[50506]: pgmap v1688: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:03.620 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:03 vm01.local ceph-mon[48597]: pgmap v1688: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:05.505 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:17:05.505 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:17:05.532 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:17:05.533 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:17:06.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:05 vm08.local ceph-mon[50506]: pgmap v1689: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:06.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:05 vm01.local ceph-mon[48597]: pgmap v1689: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:08.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:07 vm08.local ceph-mon[50506]: pgmap v1690: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:17:08.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:07 vm01.local ceph-mon[48597]: pgmap v1690: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:17:10.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:09 vm08.local ceph-mon[50506]: pgmap v1691: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:10.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:09 vm01.local ceph-mon[48597]: pgmap v1691: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:10.535 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:17:10.535 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:17:10.561 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:17:10.562 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:17:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:17:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:17:11.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:17:11.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:17:12.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:11 vm08.local ceph-mon[50506]: pgmap v1692: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:17:12.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:11 vm01.local ceph-mon[48597]: pgmap v1692: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:17:13.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:13 vm08.local ceph-mon[50506]: pgmap v1693: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:13.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:13 vm01.local ceph-mon[48597]: pgmap v1693: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:15.564 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:17:15.564 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:17:15.591 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:17:15.591 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:17:16.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:15 vm08.local ceph-mon[50506]: pgmap v1694: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:15 vm01.local ceph-mon[48597]: pgmap v1694: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:18.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:17 vm08.local ceph-mon[50506]: pgmap v1695: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:18.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:17 vm01.local ceph-mon[48597]: pgmap v1695: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:19 vm08.local ceph-mon[50506]: pgmap v1696: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:17:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:19 vm01.local ceph-mon[48597]: pgmap v1696: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:17:20.592 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:17:20.593 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:17:20.618 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:17:20.619 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:17:22.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:21 vm08.local ceph-mon[50506]: pgmap v1697: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:21 vm01.local ceph-mon[48597]: pgmap v1697: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:23 vm08.local ceph-mon[50506]: pgmap v1698: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:17:23.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:23 vm01.local ceph-mon[48597]: pgmap v1698: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:17:25.620 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:17:25.621 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:17:25.647 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:17:25.647 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:17:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:25 vm08.local ceph-mon[50506]: pgmap v1699: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:17:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:25 vm01.local ceph-mon[48597]: pgmap v1699: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:17:28.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:27 vm08.local ceph-mon[50506]: pgmap v1700: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:28.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:27 vm01.local ceph-mon[48597]: pgmap v1700: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:28.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:28 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:17:28.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:28 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:17:28.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:28 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:17:28.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:28 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:17:28.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:28 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:17:28.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:28 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:17:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:29 vm08.local ceph-mon[50506]: pgmap v1701: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:29 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:17:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:29 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:17:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:29 vm01.local ceph-mon[48597]: pgmap v1701: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:29 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:17:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:29 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:17:30.649 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:17:30.649 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:17:30.675 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:17:30.675 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:17:32.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:31 vm08.local ceph-mon[50506]: pgmap v1702: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:17:32.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:31 vm01.local ceph-mon[48597]: pgmap v1702: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:17:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:33 vm08.local ceph-mon[50506]: pgmap v1703: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:33.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:33 vm01.local ceph-mon[48597]: pgmap v1703: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:35.677 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:17:35.677 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:17:35.704 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:17:35.705 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:17:36.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:35 vm08.local ceph-mon[50506]: pgmap v1704: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:36.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:35 vm01.local ceph-mon[48597]: pgmap v1704: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:37.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:36 vm08.local ceph-mon[50506]: pgmap v1705: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:17:37.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:36 vm01.local ceph-mon[48597]: pgmap v1705: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:17:40.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:39 vm08.local ceph-mon[50506]: pgmap v1706: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:39 vm01.local ceph-mon[48597]: pgmap v1706: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:40.706 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:17:40.706 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:17:40.735 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:17:40.735 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:17:42.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:41 vm08.local ceph-mon[50506]: pgmap v1707: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:17:42.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:41 vm01.local ceph-mon[48597]: pgmap v1707: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:17:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:43 vm08.local ceph-mon[50506]: pgmap v1708: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:43.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:43 vm01.local ceph-mon[48597]: pgmap v1708: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:45.736 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:17:45.737 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:17:45.763 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:17:45.763 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:17:46.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:45 vm08.local ceph-mon[50506]: pgmap v1709: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:46.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:45 vm01.local ceph-mon[48597]: pgmap v1709: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:48.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:47 vm08.local ceph-mon[50506]: pgmap v1710: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:17:48.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:47 vm01.local ceph-mon[48597]: pgmap v1710: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:17:50.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:49 vm08.local ceph-mon[50506]: pgmap v1711: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:50.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:49 vm01.local ceph-mon[48597]: pgmap v1711: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:50.764 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:17:50.765 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:17:50.793 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:17:50.794 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:17:52.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:51 vm08.local ceph-mon[50506]: pgmap v1712: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:17:52.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:51 vm01.local ceph-mon[48597]: pgmap v1712: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:17:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:53 vm08.local ceph-mon[50506]: pgmap v1713: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:53.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:53 vm01.local ceph-mon[48597]: pgmap v1713: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:55.795 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:17:55.795 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:17:55.821 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:17:55.822 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:17:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:55 vm08.local ceph-mon[50506]: pgmap v1714: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:56.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:55 vm01.local ceph-mon[48597]: pgmap v1714: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:17:58.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:57 vm08.local ceph-mon[50506]: pgmap v1715: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:17:58.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:57 vm01.local ceph-mon[48597]: pgmap v1715: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:18:00.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:17:59 vm08.local ceph-mon[50506]: pgmap v1716: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:00.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:17:59 vm01.local ceph-mon[48597]: pgmap v1716: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:00.823 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:18:00.823 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:18:00.849 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:18:00.849 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:18:02.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:01 vm08.local ceph-mon[50506]: pgmap v1717: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:18:02.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:01 vm01.local ceph-mon[48597]: pgmap v1717: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:18:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:03 vm08.local ceph-mon[50506]: pgmap v1718: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:03.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:03 vm01.local ceph-mon[48597]: pgmap v1718: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:05.851 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:18:05.851 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:18:05.876 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:18:05.877 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:18:06.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:05 vm08.local ceph-mon[50506]: pgmap v1719: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:06.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:05 vm01.local ceph-mon[48597]: pgmap v1719: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:08.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:07 vm08.local ceph-mon[50506]: pgmap v1720: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:18:08.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:07 vm01.local ceph-mon[48597]: pgmap v1720: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:18:10.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:09 vm08.local ceph-mon[50506]: pgmap v1721: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:10.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:09 vm01.local ceph-mon[48597]: pgmap v1721: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:10.878 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:18:10.879 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:18:10.905 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:18:10.906 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:18:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:18:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:18:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:18:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:18:12.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:12 vm08.local ceph-mon[50506]: pgmap v1722: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:18:12.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:12 vm01.local ceph-mon[48597]: pgmap v1722: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:18:13.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:13 vm08.local ceph-mon[50506]: pgmap v1723: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:13.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:13 vm01.local ceph-mon[48597]: pgmap v1723: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:15.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:14 vm08.local ceph-mon[50506]: pgmap v1724: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:15.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:14 vm01.local ceph-mon[48597]: pgmap v1724: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:15.908 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:18:15.908 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:18:15.936 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:18:15.936 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:18:18.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:17 vm08.local ceph-mon[50506]: pgmap v1725: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:18:18.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:17 vm01.local ceph-mon[48597]: pgmap v1725: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:18:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:19 vm08.local ceph-mon[50506]: pgmap v1726: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:19 vm01.local ceph-mon[48597]: pgmap v1726: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:20.938 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:18:20.938 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:18:20.967 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:18:20.968 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:18:22.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:21 vm08.local ceph-mon[50506]: pgmap v1727: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:18:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:21 vm01.local ceph-mon[48597]: pgmap v1727: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:18:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:23 vm08.local ceph-mon[50506]: pgmap v1728: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:23.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:23 vm01.local ceph-mon[48597]: pgmap v1728: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:25.969 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:18:25.969 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:18:25.995 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:18:25.996 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:18:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:25 vm08.local ceph-mon[50506]: pgmap v1729: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:25 vm01.local ceph-mon[48597]: pgmap v1729: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:28.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:27 vm08.local ceph-mon[50506]: pgmap v1730: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:18:28.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:27 vm01.local ceph-mon[48597]: pgmap v1730: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:18:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:29 vm08.local ceph-mon[50506]: pgmap v1731: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:29 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:18:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:29 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:18:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:29 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:18:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:29 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:18:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:29 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:18:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:29 vm01.local ceph-mon[48597]: pgmap v1731: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:29 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:18:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:29 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:18:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:29 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:18:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:29 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:18:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:29 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:18:30.997 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:18:30.997 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:18:31.030 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:18:31.031 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:18:32.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:31 vm08.local ceph-mon[50506]: pgmap v1732: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:18:32.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:31 vm01.local ceph-mon[48597]: pgmap v1732: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:18:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:33 vm08.local ceph-mon[50506]: pgmap v1733: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:33.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:33 vm01.local ceph-mon[48597]: pgmap v1733: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:36.033 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:18:36.034 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:18:36.061 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:18:36.061 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:18:36.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:35 vm08.local ceph-mon[50506]: pgmap v1734: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:35 vm01.local ceph-mon[48597]: pgmap v1734: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:38.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:37 vm08.local ceph-mon[50506]: pgmap v1735: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:18:38.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:37 vm01.local ceph-mon[48597]: pgmap v1735: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:18:40.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:39 vm08.local ceph-mon[50506]: pgmap v1736: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:39 vm01.local ceph-mon[48597]: pgmap v1736: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:41.063 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:18:41.063 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:18:41.230 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:18:41.239 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:18:42.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:41 vm08.local ceph-mon[50506]: pgmap v1737: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:18:42.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:41 vm01.local ceph-mon[48597]: pgmap v1737: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:18:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:43 vm08.local ceph-mon[50506]: pgmap v1738: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:43.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:43 vm01.local ceph-mon[48597]: pgmap v1738: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:46.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:45 vm08.local ceph-mon[50506]: pgmap v1739: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:46.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:45 vm01.local ceph-mon[48597]: pgmap v1739: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:46.232 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:18:46.232 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:18:46.257 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:18:46.258 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:18:48.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:47 vm08.local ceph-mon[50506]: pgmap v1740: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:18:48.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:47 vm01.local ceph-mon[48597]: pgmap v1740: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:18:50.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:49 vm08.local ceph-mon[50506]: pgmap v1741: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:50.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:49 vm01.local ceph-mon[48597]: pgmap v1741: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:51.259 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:18:51.260 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:18:51.288 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:18:51.288 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:18:52.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:51 vm08.local ceph-mon[50506]: pgmap v1742: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:18:52.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:51 vm01.local ceph-mon[48597]: pgmap v1742: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:18:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:53 vm08.local ceph-mon[50506]: pgmap v1743: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:53.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:53 vm01.local ceph-mon[48597]: pgmap v1743: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:55 vm08.local ceph-mon[50506]: pgmap v1744: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:56.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:55 vm01.local ceph-mon[48597]: pgmap v1744: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:18:56.290 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:18:56.290 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:18:56.315 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:18:56.316 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:18:58.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:57 vm08.local ceph-mon[50506]: pgmap v1745: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:18:58.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:57 vm01.local ceph-mon[48597]: pgmap v1745: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:19:00.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:18:59 vm08.local ceph-mon[50506]: pgmap v1746: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:00.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:18:59 vm01.local ceph-mon[48597]: pgmap v1746: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:01.317 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:19:01.318 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:19:01.343 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:19:01.344 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:19:02.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:01 vm08.local ceph-mon[50506]: pgmap v1747: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:19:02.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:01 vm01.local ceph-mon[48597]: pgmap v1747: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:19:03.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:03 vm01.local ceph-mon[48597]: pgmap v1748: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:04.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:03 vm08.local ceph-mon[50506]: pgmap v1748: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:06.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:05 vm08.local ceph-mon[50506]: pgmap v1749: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:06.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:05 vm01.local ceph-mon[48597]: pgmap v1749: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:06.345 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:19:06.345 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:19:06.372 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:19:06.373 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:19:08.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:07 vm01.local ceph-mon[48597]: pgmap v1750: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:19:08.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:07 vm08.local ceph-mon[50506]: pgmap v1750: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:19:10.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:09 vm01.local ceph-mon[48597]: pgmap v1751: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:10.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:09 vm08.local ceph-mon[50506]: pgmap v1751: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:11.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:19:11.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:19:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:19:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:19:11.374 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:19:11.374 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:19:11.401 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:19:11.401 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:19:12.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:11 vm01.local ceph-mon[48597]: pgmap v1752: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:19:12.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:11 vm08.local ceph-mon[50506]: pgmap v1752: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:19:13.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:13 vm01.local ceph-mon[48597]: pgmap v1753: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:13 vm08.local ceph-mon[50506]: pgmap v1753: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:16.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:15 vm08.local ceph-mon[50506]: pgmap v1754: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:15 vm01.local ceph-mon[48597]: pgmap v1754: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:16.403 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:19:16.403 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:19:16.432 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:19:16.433 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:19:18.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:17 vm08.local ceph-mon[50506]: pgmap v1755: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:19:18.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:17 vm01.local ceph-mon[48597]: pgmap v1755: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:19:20.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:19 vm08.local ceph-mon[50506]: pgmap v1756: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:20.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:19 vm01.local ceph-mon[48597]: pgmap v1756: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:21 vm08.local ceph-mon[50506]: pgmap v1757: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:19:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:21 vm01.local ceph-mon[48597]: pgmap v1757: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:19:21.434 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:19:21.435 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:19:21.462 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:19:21.462 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:19:23.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:23 vm01.local ceph-mon[48597]: pgmap v1758: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:23.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:23 vm08.local ceph-mon[50506]: pgmap v1758: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:25 vm08.local ceph-mon[50506]: pgmap v1759: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:25 vm01.local ceph-mon[48597]: pgmap v1759: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:26.463 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:19:26.464 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:19:26.492 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:19:26.492 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:19:28.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:27 vm08.local ceph-mon[50506]: pgmap v1760: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:19:28.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:27 vm01.local ceph-mon[48597]: pgmap v1760: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:19:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:29 vm08.local ceph-mon[50506]: pgmap v1761: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:29 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:19:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:29 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:19:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:29 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:19:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:29 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:19:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:29 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:19:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:29 vm01.local ceph-mon[48597]: pgmap v1761: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:29 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:19:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:29 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:19:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:29 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:19:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:29 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:19:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:29 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:19:31.494 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:19:31.494 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:19:31.520 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:19:31.521 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:19:32.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:31 vm08.local ceph-mon[50506]: pgmap v1762: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:19:32.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:31 vm01.local ceph-mon[48597]: pgmap v1762: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:19:33.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:33 vm01.local ceph-mon[48597]: pgmap v1763: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:33.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:33 vm08.local ceph-mon[50506]: pgmap v1763: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:36.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:35 vm08.local ceph-mon[50506]: pgmap v1764: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:35 vm01.local ceph-mon[48597]: pgmap v1764: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:36.522 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:19:36.523 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:19:36.549 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:19:36.550 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:19:38.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:37 vm08.local ceph-mon[50506]: pgmap v1765: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:19:38.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:37 vm01.local ceph-mon[48597]: pgmap v1765: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:19:40.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:39 vm08.local ceph-mon[50506]: pgmap v1766: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:39 vm01.local ceph-mon[48597]: pgmap v1766: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:41.551 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:19:41.552 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:19:41.578 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:19:41.579 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:19:42.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:41 vm08.local ceph-mon[50506]: pgmap v1767: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:19:42.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:41 vm01.local ceph-mon[48597]: pgmap v1767: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:19:43.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:43 vm01.local ceph-mon[48597]: pgmap v1768: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:43.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:43 vm08.local ceph-mon[50506]: pgmap v1768: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:46.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:45 vm08.local ceph-mon[50506]: pgmap v1769: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:46.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:45 vm01.local ceph-mon[48597]: pgmap v1769: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:46.581 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:19:46.581 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:19:46.634 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:19:46.635 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:19:48.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:47 vm08.local ceph-mon[50506]: pgmap v1770: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:19:48.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:47 vm01.local ceph-mon[48597]: pgmap v1770: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:19:50.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:49 vm08.local ceph-mon[50506]: pgmap v1771: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:50.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:49 vm01.local ceph-mon[48597]: pgmap v1771: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:51.636 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:19:51.637 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:19:51.777 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:19:51.778 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:19:52.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:51 vm01.local ceph-mon[48597]: pgmap v1772: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:19:52.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:51 vm08.local ceph-mon[50506]: pgmap v1772: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:19:53.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:53 vm01.local ceph-mon[48597]: pgmap v1773: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:53.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:53 vm08.local ceph-mon[50506]: pgmap v1773: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:55 vm08.local ceph-mon[50506]: pgmap v1774: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:56.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:55 vm01.local ceph-mon[48597]: pgmap v1774: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:19:56.779 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:19:56.779 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:19:56.811 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:19:56.811 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:19:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:57 vm08.local ceph-mon[50506]: pgmap v1775: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:19:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:57 vm01.local ceph-mon[48597]: pgmap v1775: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:20:00.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:19:59 vm08.local ceph-mon[50506]: pgmap v1776: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:00.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:19:59 vm01.local ceph-mon[48597]: pgmap v1776: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:01.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:00 vm08.local ceph-mon[50506]: overall HEALTH_OK 2026-03-10T11:20:01.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:00 vm08.local ceph-mon[50506]: pgmap v1777: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:20:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:00 vm01.local ceph-mon[48597]: overall HEALTH_OK 2026-03-10T11:20:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:00 vm01.local ceph-mon[48597]: pgmap v1777: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:20:01.812 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:20:01.813 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:20:01.838 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:20:01.839 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:20:03.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:03 vm01.local ceph-mon[48597]: pgmap v1778: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:03.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:03 vm08.local ceph-mon[50506]: pgmap v1778: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:06.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:05 vm01.local ceph-mon[48597]: pgmap v1779: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:06.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:05 vm08.local ceph-mon[50506]: pgmap v1779: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:06.840 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:20:06.840 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:20:06.867 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:20:06.867 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:20:08.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:07 vm01.local ceph-mon[48597]: pgmap v1780: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:20:08.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:07 vm08.local ceph-mon[50506]: pgmap v1780: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:20:10.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:09 vm08.local ceph-mon[50506]: pgmap v1781: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:10.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:09 vm01.local ceph-mon[48597]: pgmap v1781: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:20:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:20:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:20:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:20:11.869 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:20:11.869 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:20:11.898 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:20:11.899 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:20:12.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:11 vm08.local ceph-mon[50506]: pgmap v1782: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:20:12.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:11 vm01.local ceph-mon[48597]: pgmap v1782: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:20:13.787 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:13 vm01.local ceph-mon[48597]: pgmap v1783: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:13 vm08.local ceph-mon[50506]: pgmap v1783: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:16.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:15 vm08.local ceph-mon[50506]: pgmap v1784: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:15 vm01.local ceph-mon[48597]: pgmap v1784: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:16.901 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:20:16.902 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:20:16.938 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:20:16.938 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:20:18.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:17 vm08.local ceph-mon[50506]: pgmap v1785: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:20:18.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:17 vm01.local ceph-mon[48597]: pgmap v1785: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:20:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:19 vm08.local ceph-mon[50506]: pgmap v1786: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:19 vm01.local ceph-mon[48597]: pgmap v1786: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:21.940 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:20:21.941 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:20:21.967 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:20:21.968 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:20:22.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:21 vm01.local ceph-mon[48597]: pgmap v1787: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:20:22.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:21 vm08.local ceph-mon[50506]: pgmap v1787: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:20:23.789 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:23 vm01.local ceph-mon[48597]: pgmap v1788: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:23.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:23 vm08.local ceph-mon[50506]: pgmap v1788: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:25 vm08.local ceph-mon[50506]: pgmap v1789: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:25 vm01.local ceph-mon[48597]: pgmap v1789: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:26.969 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:20:26.970 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:20:26.997 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:20:26.997 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:20:28.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:27 vm08.local ceph-mon[50506]: pgmap v1790: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:20:28.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:27 vm01.local ceph-mon[48597]: pgmap v1790: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:20:30.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:29 vm08.local ceph-mon[50506]: pgmap v1791: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:30.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:29 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:20:30.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:29 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:20:30.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:29 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:20:30.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:29 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:20:30.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:29 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:20:30.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:29 vm01.local ceph-mon[48597]: pgmap v1791: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:30.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:29 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:20:30.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:29 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:20:30.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:29 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:20:30.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:29 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:20:30.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:29 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:20:31.999 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:20:31.999 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:20:32.026 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:20:32.027 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:20:32.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:31 vm08.local ceph-mon[50506]: pgmap v1792: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:20:32.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:31 vm01.local ceph-mon[48597]: pgmap v1792: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:20:33.790 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:33 vm01.local ceph-mon[48597]: pgmap v1793: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:33.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:33 vm08.local ceph-mon[50506]: pgmap v1793: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:36.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:35 vm08.local ceph-mon[50506]: pgmap v1794: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:35 vm01.local ceph-mon[48597]: pgmap v1794: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:37.028 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:20:37.029 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:20:37.055 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:20:37.055 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:20:38.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:37 vm08.local ceph-mon[50506]: pgmap v1795: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:20:38.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:37 vm01.local ceph-mon[48597]: pgmap v1795: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:20:40.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:39 vm08.local ceph-mon[50506]: pgmap v1796: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:39 vm01.local ceph-mon[48597]: pgmap v1796: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:42.056 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:20:42.057 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:20:42.088 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:20:42.088 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:20:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:41 vm08.local ceph-mon[50506]: pgmap v1797: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:20:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:41 vm01.local ceph-mon[48597]: pgmap v1797: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:20:43.791 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:43 vm01.local ceph-mon[48597]: pgmap v1798: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:43.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:43 vm08.local ceph-mon[50506]: pgmap v1798: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:46.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:45 vm08.local ceph-mon[50506]: pgmap v1799: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:46.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:45 vm01.local ceph-mon[48597]: pgmap v1799: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:47.090 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:20:47.090 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:20:47.116 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:20:47.116 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:20:48.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:47 vm08.local ceph-mon[50506]: pgmap v1800: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:20:48.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:47 vm01.local ceph-mon[48597]: pgmap v1800: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:20:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:49 vm08.local ceph-mon[50506]: pgmap v1801: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:49.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:49 vm01.local ceph-mon[48597]: pgmap v1801: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:52.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:51 vm08.local ceph-mon[50506]: pgmap v1802: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:20:52.117 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:20:52.118 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:20:52.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:51 vm01.local ceph-mon[48597]: pgmap v1802: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:20:52.143 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:20:52.144 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:20:53.792 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:53 vm01.local ceph-mon[48597]: pgmap v1803: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:53.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:53 vm08.local ceph-mon[50506]: pgmap v1803: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:55 vm08.local ceph-mon[50506]: pgmap v1804: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:56.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:55 vm01.local ceph-mon[48597]: pgmap v1804: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:57.145 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:20:57.146 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:20:57.212 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:20:57.213 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:20:58.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:57 vm01.local ceph-mon[48597]: pgmap v1805: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:20:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:57 vm08.local ceph-mon[50506]: pgmap v1805: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:20:59.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:20:58 vm08.local ceph-mon[50506]: pgmap v1806: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:20:59.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:20:58 vm01.local ceph-mon[48597]: pgmap v1806: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:02.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:01 vm08.local ceph-mon[50506]: pgmap v1807: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:21:02.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:01 vm01.local ceph-mon[48597]: pgmap v1807: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:21:02.214 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:21:02.215 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:21:02.243 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:21:02.243 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:21:03.793 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:03 vm01.local ceph-mon[48597]: pgmap v1808: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:03.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:03 vm08.local ceph-mon[50506]: pgmap v1808: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:06.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:05 vm08.local ceph-mon[50506]: pgmap v1809: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:06.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:05 vm01.local ceph-mon[48597]: pgmap v1809: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:07.244 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:21:07.245 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:21:07.273 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:21:07.273 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:21:07.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:07 vm01.local ceph-mon[48597]: pgmap v1810: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:21:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:07 vm08.local ceph-mon[50506]: pgmap v1810: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:21:10.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:10 vm08.local ceph-mon[50506]: pgmap v1811: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:10.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:09 vm01.local ceph-mon[48597]: pgmap v1811: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:21:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:21:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:10 vm08.local ceph-mon[50506]: pgmap v1812: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:21:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:21:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:21:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:10 vm01.local ceph-mon[48597]: pgmap v1812: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:21:12.275 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:21:12.275 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:21:12.325 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:21:12.326 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:21:13.794 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:13 vm01.local ceph-mon[48597]: pgmap v1813: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:13 vm08.local ceph-mon[50506]: pgmap v1813: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:16.090 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:15 vm01.local ceph-mon[48597]: pgmap v1814: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:16.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:15 vm08.local ceph-mon[50506]: pgmap v1814: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:17.328 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:21:17.329 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:21:17.395 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:21:17.395 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:21:18.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:17 vm08.local ceph-mon[50506]: pgmap v1815: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:21:18.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:17 vm01.local ceph-mon[48597]: pgmap v1815: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:21:20.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:20 vm08.local ceph-mon[50506]: pgmap v1816: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:20.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:20 vm01.local ceph-mon[48597]: pgmap v1816: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:21 vm01.local ceph-mon[48597]: pgmap v1817: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:21:21.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:21 vm08.local ceph-mon[50506]: pgmap v1817: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:21:22.397 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:21:22.397 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:21:22.426 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:21:22.426 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:21:23.794 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:23 vm01.local ceph-mon[48597]: pgmap v1818: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:23.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:23 vm08.local ceph-mon[50506]: pgmap v1818: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:25 vm08.local ceph-mon[50506]: pgmap v1819: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:25 vm01.local ceph-mon[48597]: pgmap v1819: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:27.427 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:21:27.428 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:21:27.560 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:21:27.560 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:21:28.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:27 vm08.local ceph-mon[50506]: pgmap v1820: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:21:28.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:27 vm01.local ceph-mon[48597]: pgmap v1820: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:21:30.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:30 vm08.local ceph-mon[50506]: pgmap v1821: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:30.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:30 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:21:30.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:30 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:21:30.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:30 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:21:30.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:30 vm01.local ceph-mon[48597]: pgmap v1821: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:30.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:30 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:21:30.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:30 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:21:30.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:30 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:21:31.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:31 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:21:31.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:31 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:21:31.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:31 vm01.local ceph-mon[48597]: pgmap v1822: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:21:31.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:31 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:21:31.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:31 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:21:31.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:31 vm08.local ceph-mon[50506]: pgmap v1822: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:21:32.562 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:21:32.563 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:21:32.589 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:21:32.589 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:21:33.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:33 vm08.local ceph-mon[50506]: pgmap v1823: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:33.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:33 vm01.local ceph-mon[48597]: pgmap v1823: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:36.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:35 vm08.local ceph-mon[50506]: pgmap v1824: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:35 vm01.local ceph-mon[48597]: pgmap v1824: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:37.592 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:21:37.592 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:21:37.619 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:21:37.619 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:21:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:37 vm08.local ceph-mon[50506]: pgmap v1825: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:21:38.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:37 vm01.local ceph-mon[48597]: pgmap v1825: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:21:40.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:39 vm08.local ceph-mon[50506]: pgmap v1826: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:40.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:39 vm01.local ceph-mon[48597]: pgmap v1826: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:41 vm08.local ceph-mon[50506]: pgmap v1827: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:21:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:41 vm01.local ceph-mon[48597]: pgmap v1827: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:21:42.621 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:21:42.622 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:21:42.649 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:21:42.649 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:21:43.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:43 vm01.local ceph-mon[48597]: pgmap v1828: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:44.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:43 vm08.local ceph-mon[50506]: pgmap v1828: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:46.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:45 vm08.local ceph-mon[50506]: pgmap v1829: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:46.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:45 vm01.local ceph-mon[48597]: pgmap v1829: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:47.651 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:21:47.651 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:21:47.677 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:21:47.678 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:21:48.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:47 vm08.local ceph-mon[50506]: pgmap v1830: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:21:48.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:47 vm01.local ceph-mon[48597]: pgmap v1830: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:21:50.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:49 vm01.local ceph-mon[48597]: pgmap v1831: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:50.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:49 vm08.local ceph-mon[50506]: pgmap v1831: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:52.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:51 vm01.local ceph-mon[48597]: pgmap v1832: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:21:52.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:51 vm08.local ceph-mon[50506]: pgmap v1832: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:21:52.679 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:21:52.679 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:21:52.769 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:21:52.770 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:21:54.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:53 vm08.local ceph-mon[50506]: pgmap v1833: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:53 vm01.local ceph-mon[48597]: pgmap v1833: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:56.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:55 vm08.local ceph-mon[50506]: pgmap v1834: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:56.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:55 vm01.local ceph-mon[48597]: pgmap v1834: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:21:57.771 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:21:57.772 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:21:57.798 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:21:57.798 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:21:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:57 vm08.local ceph-mon[50506]: pgmap v1835: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:21:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:57 vm01.local ceph-mon[48597]: pgmap v1835: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:22:00.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:21:59 vm08.local ceph-mon[50506]: pgmap v1836: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:00.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:21:59 vm01.local ceph-mon[48597]: pgmap v1836: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:02.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:01 vm08.local ceph-mon[50506]: pgmap v1837: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:22:02.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:01 vm01.local ceph-mon[48597]: pgmap v1837: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:22:02.800 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:22:02.800 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:22:02.827 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:22:02.827 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:22:04.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:03 vm08.local ceph-mon[50506]: pgmap v1838: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:04.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:03 vm01.local ceph-mon[48597]: pgmap v1838: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:06.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:05 vm08.local ceph-mon[50506]: pgmap v1839: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:06.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:05 vm01.local ceph-mon[48597]: pgmap v1839: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:07.829 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:22:07.829 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:22:07.855 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:22:07.855 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:22:08.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:07 vm08.local ceph-mon[50506]: pgmap v1840: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:22:08.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:07 vm01.local ceph-mon[48597]: pgmap v1840: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:22:10.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:09 vm08.local ceph-mon[50506]: pgmap v1841: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:10.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:09 vm01.local ceph-mon[48597]: pgmap v1841: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:22:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:22:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:22:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:22:12.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:11 vm08.local ceph-mon[50506]: pgmap v1842: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:22:12.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:11 vm01.local ceph-mon[48597]: pgmap v1842: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:22:12.857 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:22:12.857 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:22:12.883 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:22:12.884 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:22:14.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:13 vm08.local ceph-mon[50506]: pgmap v1843: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:14.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:13 vm01.local ceph-mon[48597]: pgmap v1843: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:16.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:15 vm08.local ceph-mon[50506]: pgmap v1844: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:15 vm01.local ceph-mon[48597]: pgmap v1844: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:17.885 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:22:17.886 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:22:17.989 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:22:17.989 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:22:18.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:17 vm01.local ceph-mon[48597]: pgmap v1845: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:22:18.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:17 vm08.local ceph-mon[50506]: pgmap v1845: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:22:20.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:20 vm08.local ceph-mon[50506]: pgmap v1846: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:20.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:19 vm01.local ceph-mon[48597]: pgmap v1846: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:21 vm08.local ceph-mon[50506]: pgmap v1847: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:22:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:21 vm01.local ceph-mon[48597]: pgmap v1847: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:22:22.991 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:22:22.991 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:22:23.018 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:22:23.018 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:22:23.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:23 vm08.local ceph-mon[50506]: pgmap v1848: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:23.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:23 vm01.local ceph-mon[48597]: pgmap v1848: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:25 vm08.local ceph-mon[50506]: pgmap v1849: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:25 vm01.local ceph-mon[48597]: pgmap v1849: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:28.019 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:22:28.020 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:22:28.046 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:22:28.046 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:22:28.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:27 vm01.local ceph-mon[48597]: pgmap v1850: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:22:28.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:27 vm08.local ceph-mon[50506]: pgmap v1850: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:22:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:29 vm01.local ceph-mon[48597]: pgmap v1851: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:30.271 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:29 vm08.local ceph-mon[50506]: pgmap v1851: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:31.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:30 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:22:31.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:30 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:22:31.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:30 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:22:31.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:30 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:22:31.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:30 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:22:31.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:30 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:22:31.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:30 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:22:31.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:30 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:22:31.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:30 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:22:31.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:30 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:22:32.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:31 vm08.local ceph-mon[50506]: pgmap v1852: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:22:32.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:31 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:22:32.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:31 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:22:32.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:31 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:22:32.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:31 vm01.local ceph-mon[48597]: pgmap v1852: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:22:32.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:31 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:22:32.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:31 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:22:32.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:31 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:22:33.047 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:22:33.048 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:22:33.075 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:22:33.075 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:22:34.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:33 vm08.local ceph-mon[50506]: pgmap v1853: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:34.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:33 vm01.local ceph-mon[48597]: pgmap v1853: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:36.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:35 vm08.local ceph-mon[50506]: pgmap v1854: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:35 vm01.local ceph-mon[48597]: pgmap v1854: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:38.076 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:22:38.077 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:22:38.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:37 vm08.local ceph-mon[50506]: pgmap v1855: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:22:38.104 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:22:38.104 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:22:38.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:37 vm01.local ceph-mon[48597]: pgmap v1855: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:22:40.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:39 vm08.local ceph-mon[50506]: pgmap v1856: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:39 vm01.local ceph-mon[48597]: pgmap v1856: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:42.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:41 vm08.local ceph-mon[50506]: pgmap v1857: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:22:42.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:41 vm01.local ceph-mon[48597]: pgmap v1857: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:22:43.106 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:22:43.106 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:22:43.133 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:22:43.134 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:22:44.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:43 vm08.local ceph-mon[50506]: pgmap v1858: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:44.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:43 vm01.local ceph-mon[48597]: pgmap v1858: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:46.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:45 vm01.local ceph-mon[48597]: pgmap v1859: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:46.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:45 vm08.local ceph-mon[50506]: pgmap v1859: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:48.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:47 vm01.local ceph-mon[48597]: pgmap v1860: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:22:48.135 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:22:48.136 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:22:48.169 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:22:48.170 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:22:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:47 vm08.local ceph-mon[50506]: pgmap v1860: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:22:50.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:49 vm08.local ceph-mon[50506]: pgmap v1861: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:50.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:49 vm01.local ceph-mon[48597]: pgmap v1861: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:52.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:51 vm08.local ceph-mon[50506]: pgmap v1862: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:22:52.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:51 vm01.local ceph-mon[48597]: pgmap v1862: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:22:53.172 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:22:53.172 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:22:53.200 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:22:53.200 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:22:54.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:53 vm08.local ceph-mon[50506]: pgmap v1863: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:53 vm01.local ceph-mon[48597]: pgmap v1863: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:56.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:55 vm08.local ceph-mon[50506]: pgmap v1864: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:56.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:55 vm01.local ceph-mon[48597]: pgmap v1864: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:58.202 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:22:58.202 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:22:58.230 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:22:58.231 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:22:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:57 vm08.local ceph-mon[50506]: pgmap v1865: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:22:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:57 vm01.local ceph-mon[48597]: pgmap v1865: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:22:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:22:59 vm08.local ceph-mon[50506]: pgmap v1866: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:22:59.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:22:59 vm01.local ceph-mon[48597]: pgmap v1866: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:01 vm08.local ceph-mon[50506]: pgmap v1867: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:23:01.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:01 vm01.local ceph-mon[48597]: pgmap v1867: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:23:03.233 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:23:03.233 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:23:03.260 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:23:03.260 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:23:04.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:03 vm01.local ceph-mon[48597]: pgmap v1868: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:04.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:03 vm08.local ceph-mon[50506]: pgmap v1868: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:06.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:05 vm01.local ceph-mon[48597]: pgmap v1869: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:06.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:05 vm08.local ceph-mon[50506]: pgmap v1869: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:08.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:07 vm01.local ceph-mon[48597]: pgmap v1870: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:23:08.262 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:23:08.262 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:23:08.289 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:23:08.289 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:23:08.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:07 vm08.local ceph-mon[50506]: pgmap v1870: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:23:10.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:09 vm01.local ceph-mon[48597]: pgmap v1871: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:10.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:09 vm08.local ceph-mon[50506]: pgmap v1871: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:11.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:23:11.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:23:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:23:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:23:12.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:11 vm08.local ceph-mon[50506]: pgmap v1872: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:23:12.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:11 vm01.local ceph-mon[48597]: pgmap v1872: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:23:13.291 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:23:13.291 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:23:13.317 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:23:13.317 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:23:14.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:13 vm08.local ceph-mon[50506]: pgmap v1873: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:14.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:13 vm01.local ceph-mon[48597]: pgmap v1873: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:15 vm01.local ceph-mon[48597]: pgmap v1874: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:16.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:15 vm08.local ceph-mon[50506]: pgmap v1874: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:18.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:17 vm01.local ceph-mon[48597]: pgmap v1875: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:23:18.319 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:23:18.319 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:23:18.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:17 vm08.local ceph-mon[50506]: pgmap v1875: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:23:18.346 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:23:18.346 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:23:20.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:19 vm08.local ceph-mon[50506]: pgmap v1876: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:20.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:19 vm01.local ceph-mon[48597]: pgmap v1876: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:22.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:21 vm08.local ceph-mon[50506]: pgmap v1877: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:23:22.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:21 vm01.local ceph-mon[48597]: pgmap v1877: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:23:23.347 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:23:23.348 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:23:23.502 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:23:23.503 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:23:24.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:23 vm01.local ceph-mon[48597]: pgmap v1878: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:24.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:23 vm08.local ceph-mon[50506]: pgmap v1878: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:26.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:25 vm08.local ceph-mon[50506]: pgmap v1879: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:26.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:25 vm01.local ceph-mon[48597]: pgmap v1879: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:28.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:27 vm08.local ceph-mon[50506]: pgmap v1880: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:23:28.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:27 vm01.local ceph-mon[48597]: pgmap v1880: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:23:28.504 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:23:28.505 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:23:28.531 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:23:28.531 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:23:30.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:29 vm08.local ceph-mon[50506]: pgmap v1881: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:30.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:29 vm01.local ceph-mon[48597]: pgmap v1881: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:32.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:31 vm08.local ceph-mon[50506]: pgmap v1882: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:23:32.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:31 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:23:32.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:31 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:23:32.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:31 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:23:32.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:31 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:23:32.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:31 vm01.local ceph-mon[48597]: pgmap v1882: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:23:32.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:31 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:23:32.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:31 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:23:32.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:31 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:23:32.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:31 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:23:33.533 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:23:33.534 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:23:33.561 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:23:33.562 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:23:33.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:33 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:23:33.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:33 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:23:33.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:33 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:23:33.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:33 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:23:33.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:33 vm08.local ceph-mon[50506]: pgmap v1883: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:33.853 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:33 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:23:33.853 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:33 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:23:33.853 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:33 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:23:33.853 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:33 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:23:33.853 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:33 vm01.local ceph-mon[48597]: pgmap v1883: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:35 vm01.local ceph-mon[48597]: pgmap v1884: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:36.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:35 vm08.local ceph-mon[50506]: pgmap v1884: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:38.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:37 vm01.local ceph-mon[48597]: pgmap v1885: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:23:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:37 vm08.local ceph-mon[50506]: pgmap v1885: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:23:38.563 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:23:38.564 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:23:38.594 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:23:38.594 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:23:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:39 vm01.local ceph-mon[48597]: pgmap v1886: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:40.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:39 vm08.local ceph-mon[50506]: pgmap v1886: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:41 vm08.local ceph-mon[50506]: pgmap v1887: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:23:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:41 vm01.local ceph-mon[48597]: pgmap v1887: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:23:43.596 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:23:43.596 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:23:43.622 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:23:43.623 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:23:44.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:43 vm01.local ceph-mon[48597]: pgmap v1888: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:44.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:43 vm08.local ceph-mon[50506]: pgmap v1888: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:46.146 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:45 vm08.local ceph-mon[50506]: pgmap v1889: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:46.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:45 vm01.local ceph-mon[48597]: pgmap v1889: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:48.170 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:47 vm01.local ceph-mon[48597]: pgmap v1890: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:23:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:47 vm08.local ceph-mon[50506]: pgmap v1890: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:23:48.624 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:23:48.625 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:23:48.658 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:23:48.659 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:23:50.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:49 vm08.local ceph-mon[50506]: pgmap v1891: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:50.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:49 vm01.local ceph-mon[48597]: pgmap v1891: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:52.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:51 vm08.local ceph-mon[50506]: pgmap v1892: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:23:52.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:51 vm01.local ceph-mon[48597]: pgmap v1892: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:23:53.661 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:23:53.661 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:23:53.688 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:23:53.689 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:23:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:53 vm01.local ceph-mon[48597]: pgmap v1893: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:54.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:53 vm08.local ceph-mon[50506]: pgmap v1893: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:56.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:55 vm08.local ceph-mon[50506]: pgmap v1894: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:56.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:55 vm01.local ceph-mon[48597]: pgmap v1894: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:23:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:57 vm08.local ceph-mon[50506]: pgmap v1895: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:23:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:57 vm01.local ceph-mon[48597]: pgmap v1895: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:23:58.691 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:23:58.691 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:23:58.718 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:23:58.718 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:24:00.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:23:59 vm08.local ceph-mon[50506]: pgmap v1896: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:00.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:23:59 vm01.local ceph-mon[48597]: pgmap v1896: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:02.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:01 vm08.local ceph-mon[50506]: pgmap v1897: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:24:02.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:01 vm01.local ceph-mon[48597]: pgmap v1897: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:24:03.720 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:24:03.720 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:24:03.746 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:24:03.746 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:24:04.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:03 vm01.local ceph-mon[48597]: pgmap v1898: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:04.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:03 vm08.local ceph-mon[50506]: pgmap v1898: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:06.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:06 vm08.local ceph-mon[50506]: pgmap v1899: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:06.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:06 vm01.local ceph-mon[48597]: pgmap v1899: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:07.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:07 vm08.local ceph-mon[50506]: pgmap v1900: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:24:07.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:07 vm01.local ceph-mon[48597]: pgmap v1900: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:24:08.748 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:24:08.748 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:24:08.775 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:24:08.775 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:24:10.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:09 vm08.local ceph-mon[50506]: pgmap v1901: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:10.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:09 vm01.local ceph-mon[48597]: pgmap v1901: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:24:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:24:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:24:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:24:12.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:11 vm08.local ceph-mon[50506]: pgmap v1902: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:24:12.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:11 vm01.local ceph-mon[48597]: pgmap v1902: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:24:13.777 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:24:13.777 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:24:13.804 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:24:13.804 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:24:14.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:13 vm01.local ceph-mon[48597]: pgmap v1903: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:14.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:13 vm08.local ceph-mon[50506]: pgmap v1903: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:16.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:15 vm08.local ceph-mon[50506]: pgmap v1904: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:16.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:15 vm01.local ceph-mon[48597]: pgmap v1904: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:18.170 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:17 vm01.local ceph-mon[48597]: pgmap v1905: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:24:18.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:17 vm08.local ceph-mon[50506]: pgmap v1905: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:24:18.806 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:24:18.806 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:24:18.835 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:24:18.835 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:24:20.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:19 vm08.local ceph-mon[50506]: pgmap v1906: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:20.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:19 vm01.local ceph-mon[48597]: pgmap v1906: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:22.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:21 vm08.local ceph-mon[50506]: pgmap v1907: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:24:22.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:21 vm01.local ceph-mon[48597]: pgmap v1907: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:24:23.837 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:24:23.837 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:24:23.939 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:24:23.940 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:24:24.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:23 vm08.local ceph-mon[50506]: pgmap v1908: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:24.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:23 vm01.local ceph-mon[48597]: pgmap v1908: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:26.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:25 vm08.local ceph-mon[50506]: pgmap v1909: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:26.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:25 vm01.local ceph-mon[48597]: pgmap v1909: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:28.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:27 vm01.local ceph-mon[48597]: pgmap v1910: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:24:28.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:27 vm08.local ceph-mon[50506]: pgmap v1910: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:24:28.941 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:24:28.941 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:24:28.967 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:24:28.968 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:24:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:29 vm01.local ceph-mon[48597]: pgmap v1911: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:30.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:29 vm08.local ceph-mon[50506]: pgmap v1911: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:32.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:31 vm08.local ceph-mon[50506]: pgmap v1912: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:24:32.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:31 vm01.local ceph-mon[48597]: pgmap v1912: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:24:32.880 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:32 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:24:32.881 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:32 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:24:32.881 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:32 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:24:32.881 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:32 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:24:32.881 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:32 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:24:32.881 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:32 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:24:33.971 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:24:33.977 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:24:34.009 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:24:34.010 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:24:34.014 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:33 vm08.local ceph-mon[50506]: pgmap v1913: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:34.014 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:33 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:24:34.068 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:33 vm01.local ceph-mon[48597]: pgmap v1913: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:34.068 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:33 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:24:35.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:34 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:24:35.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:34 vm08.local ceph-mon[50506]: pgmap v1914: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:35.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:34 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:24:35.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:34 vm01.local ceph-mon[48597]: pgmap v1914: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:36.436 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:36 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:24:36.436 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:36 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:24:36.436 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:36 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:24:36.436 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:36 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:24:36.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:36 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:24:36.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:36 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:24:36.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:36 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:24:36.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:36 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:24:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:37 vm08.local ceph-mon[50506]: pgmap v1915: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:24:37.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:37 vm01.local ceph-mon[48597]: pgmap v1915: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:24:39.011 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:24:39.012 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:24:39.039 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:24:39.040 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:24:40.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:39 vm08.local ceph-mon[50506]: pgmap v1916: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:39 vm01.local ceph-mon[48597]: pgmap v1916: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:42.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:41 vm08.local ceph-mon[50506]: pgmap v1917: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:24:42.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:41 vm01.local ceph-mon[48597]: pgmap v1917: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:24:44.041 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:24:44.042 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:24:44.074 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:24:44.074 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:24:44.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:43 vm08.local ceph-mon[50506]: pgmap v1918: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:44.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:43 vm01.local ceph-mon[48597]: pgmap v1918: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:46.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:45 vm08.local ceph-mon[50506]: pgmap v1919: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:46.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:45 vm01.local ceph-mon[48597]: pgmap v1919: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:48.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:47 vm01.local ceph-mon[48597]: pgmap v1920: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:24:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:47 vm08.local ceph-mon[50506]: pgmap v1920: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:24:49.076 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:24:49.076 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:24:49.102 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:24:49.103 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:24:50.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:49 vm01.local ceph-mon[48597]: pgmap v1921: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:50.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:49 vm08.local ceph-mon[50506]: pgmap v1921: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:52.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:51 vm01.local ceph-mon[48597]: pgmap v1922: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:24:52.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:51 vm08.local ceph-mon[50506]: pgmap v1922: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:24:54.105 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:24:54.105 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:24:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:53 vm01.local ceph-mon[48597]: pgmap v1923: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:54.132 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:24:54.133 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:24:54.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:53 vm08.local ceph-mon[50506]: pgmap v1923: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:56.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:55 vm08.local ceph-mon[50506]: pgmap v1924: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:56.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:55 vm01.local ceph-mon[48597]: pgmap v1924: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:24:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:57 vm08.local ceph-mon[50506]: pgmap v1925: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:24:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:57 vm01.local ceph-mon[48597]: pgmap v1925: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:24:59.135 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:24:59.135 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:24:59.160 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:24:59.161 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:25:00.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:24:59 vm08.local ceph-mon[50506]: pgmap v1926: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:00.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:24:59 vm01.local ceph-mon[48597]: pgmap v1926: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:02.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:01 vm08.local ceph-mon[50506]: pgmap v1927: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:25:02.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:01 vm01.local ceph-mon[48597]: pgmap v1927: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:25:04.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:03 vm01.local ceph-mon[48597]: pgmap v1928: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:04.162 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:25:04.163 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:25:04.188 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:25:04.189 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:25:04.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:03 vm08.local ceph-mon[50506]: pgmap v1928: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:06.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:05 vm08.local ceph-mon[50506]: pgmap v1929: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:06.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:05 vm01.local ceph-mon[48597]: pgmap v1929: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:08.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:07 vm08.local ceph-mon[50506]: pgmap v1930: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:25:08.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:07 vm01.local ceph-mon[48597]: pgmap v1930: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:25:09.190 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:25:09.191 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:25:09.217 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:25:09.217 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:25:10.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:09 vm08.local ceph-mon[50506]: pgmap v1931: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:10.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:09 vm01.local ceph-mon[48597]: pgmap v1931: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:25:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:25:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:25:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:25:12.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:11 vm08.local ceph-mon[50506]: pgmap v1932: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:25:12.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:11 vm01.local ceph-mon[48597]: pgmap v1932: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:25:14.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:13 vm01.local ceph-mon[48597]: pgmap v1933: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:14.219 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:25:14.219 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:25:14.249 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:25:14.249 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:25:14.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:13 vm08.local ceph-mon[50506]: pgmap v1933: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:16.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:15 vm08.local ceph-mon[50506]: pgmap v1934: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:16.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:15 vm01.local ceph-mon[48597]: pgmap v1934: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:18.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:17 vm08.local ceph-mon[50506]: pgmap v1935: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:25:18.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:17 vm01.local ceph-mon[48597]: pgmap v1935: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:25:19.251 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:25:19.251 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:25:19.280 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:25:19.281 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:25:20.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:19 vm08.local ceph-mon[50506]: pgmap v1936: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:20.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:19 vm01.local ceph-mon[48597]: pgmap v1936: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:22.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:21 vm08.local ceph-mon[50506]: pgmap v1937: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:25:22.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:21 vm01.local ceph-mon[48597]: pgmap v1937: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:25:24.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:23 vm01.local ceph-mon[48597]: pgmap v1938: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:24.282 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:25:24.283 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:25:24.310 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:25:24.310 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:25:24.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:23 vm08.local ceph-mon[50506]: pgmap v1938: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:26.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:25 vm08.local ceph-mon[50506]: pgmap v1939: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:26.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:25 vm01.local ceph-mon[48597]: pgmap v1939: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:28.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:27 vm08.local ceph-mon[50506]: pgmap v1940: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:25:28.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:27 vm01.local ceph-mon[48597]: pgmap v1940: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:25:29.311 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:25:29.312 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:25:29.338 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:25:29.338 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:25:30.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:29 vm08.local ceph-mon[50506]: pgmap v1941: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:30.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:29 vm01.local ceph-mon[48597]: pgmap v1941: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:32.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:31 vm08.local ceph-mon[50506]: pgmap v1942: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:25:32.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:31 vm01.local ceph-mon[48597]: pgmap v1942: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:25:34.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:33 vm08.local ceph-mon[50506]: pgmap v1943: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:34.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:33 vm01.local ceph-mon[48597]: pgmap v1943: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:34.340 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:25:34.340 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:25:34.368 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:25:34.369 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:25:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:35 vm01.local ceph-mon[48597]: pgmap v1944: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:35 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:25:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:35 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:25:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:35 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:25:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:35 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:25:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:35 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:25:36.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:35 vm08.local ceph-mon[50506]: pgmap v1944: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:36.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:35 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:25:36.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:35 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:25:36.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:35 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:25:36.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:35 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:25:36.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:35 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:25:38.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:37 vm01.local ceph-mon[48597]: pgmap v1945: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:25:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:37 vm08.local ceph-mon[50506]: pgmap v1945: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:25:39.370 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:25:39.371 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:25:39.397 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:25:39.397 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:25:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:39 vm01.local ceph-mon[48597]: pgmap v1946: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:40.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:39 vm08.local ceph-mon[50506]: pgmap v1946: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:42.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:41 vm01.local ceph-mon[48597]: pgmap v1947: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:25:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:41 vm08.local ceph-mon[50506]: pgmap v1947: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:25:44.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:43 vm01.local ceph-mon[48597]: pgmap v1948: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:44.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:43 vm08.local ceph-mon[50506]: pgmap v1948: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:44.398 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:25:44.399 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:25:44.426 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:25:44.426 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:25:46.146 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:45 vm08.local ceph-mon[50506]: pgmap v1949: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:46.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:45 vm01.local ceph-mon[48597]: pgmap v1949: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:47 vm08.local ceph-mon[50506]: pgmap v1950: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:25:48.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:47 vm01.local ceph-mon[48597]: pgmap v1950: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:25:49.428 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:25:49.428 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:25:49.465 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:25:49.465 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:25:50.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:49 vm08.local ceph-mon[50506]: pgmap v1951: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:50.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:49 vm01.local ceph-mon[48597]: pgmap v1951: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:52.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:51 vm08.local ceph-mon[50506]: pgmap v1952: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:25:52.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:51 vm01.local ceph-mon[48597]: pgmap v1952: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:25:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:53 vm01.local ceph-mon[48597]: pgmap v1953: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:54.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:53 vm08.local ceph-mon[50506]: pgmap v1953: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:54.467 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:25:54.467 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:25:54.493 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:25:54.494 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:25:56.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:55 vm08.local ceph-mon[50506]: pgmap v1954: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:56.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:55 vm01.local ceph-mon[48597]: pgmap v1954: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:58 vm08.local ceph-mon[50506]: pgmap v1955: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:25:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:58 vm01.local ceph-mon[48597]: pgmap v1955: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:25:59.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:25:59 vm08.local ceph-mon[50506]: pgmap v1956: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:59.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:25:59 vm01.local ceph-mon[48597]: pgmap v1956: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:25:59.495 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:25:59.496 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:25:59.522 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:25:59.522 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:26:02.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:01 vm08.local ceph-mon[50506]: pgmap v1957: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:26:02.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:01 vm01.local ceph-mon[48597]: pgmap v1957: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:26:04.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:03 vm01.local ceph-mon[48597]: pgmap v1958: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:04.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:03 vm08.local ceph-mon[50506]: pgmap v1958: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:04.524 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:26:04.525 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:26:04.553 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:26:04.554 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:26:06.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:05 vm08.local ceph-mon[50506]: pgmap v1959: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:06.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:05 vm01.local ceph-mon[48597]: pgmap v1959: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:08.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:07 vm08.local ceph-mon[50506]: pgmap v1960: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:26:08.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:07 vm01.local ceph-mon[48597]: pgmap v1960: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:26:09.555 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:26:09.556 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:26:09.584 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:26:09.585 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:26:10.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:09 vm08.local ceph-mon[50506]: pgmap v1961: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:10.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:09 vm01.local ceph-mon[48597]: pgmap v1961: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:26:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:26:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:26:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:26:12.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:11 vm08.local ceph-mon[50506]: pgmap v1962: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:26:12.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:11 vm01.local ceph-mon[48597]: pgmap v1962: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:26:14.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:13 vm01.local ceph-mon[48597]: pgmap v1963: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:14.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:13 vm08.local ceph-mon[50506]: pgmap v1963: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:14.586 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:26:14.587 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:26:14.613 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:26:14.613 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:26:16.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:15 vm08.local ceph-mon[50506]: pgmap v1964: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:16.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:15 vm01.local ceph-mon[48597]: pgmap v1964: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:18.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:17 vm08.local ceph-mon[50506]: pgmap v1965: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:26:18.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:17 vm01.local ceph-mon[48597]: pgmap v1965: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:26:19.615 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:26:19.615 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:26:19.641 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:26:19.642 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:26:20.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:19 vm08.local ceph-mon[50506]: pgmap v1966: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:20.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:19 vm01.local ceph-mon[48597]: pgmap v1966: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:22.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:21 vm08.local ceph-mon[50506]: pgmap v1967: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:26:22.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:21 vm01.local ceph-mon[48597]: pgmap v1967: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:26:24.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:23 vm01.local ceph-mon[48597]: pgmap v1968: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:24.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:23 vm08.local ceph-mon[50506]: pgmap v1968: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:24.643 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:26:24.644 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:26:24.670 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:26:24.670 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:26:26.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:25 vm08.local ceph-mon[50506]: pgmap v1969: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:26.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:25 vm01.local ceph-mon[48597]: pgmap v1969: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:28.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:27 vm08.local ceph-mon[50506]: pgmap v1970: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:26:28.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:27 vm01.local ceph-mon[48597]: pgmap v1970: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:26:29.672 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:26:29.673 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:26:29.699 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:26:29.699 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:26:30.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:29 vm08.local ceph-mon[50506]: pgmap v1971: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:30.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:29 vm01.local ceph-mon[48597]: pgmap v1971: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:32.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:31 vm08.local ceph-mon[50506]: pgmap v1972: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:26:32.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:31 vm01.local ceph-mon[48597]: pgmap v1972: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:26:34.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:33 vm01.local ceph-mon[48597]: pgmap v1973: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:34.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:33 vm08.local ceph-mon[50506]: pgmap v1973: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:34.701 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:26:34.702 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:26:34.728 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:26:34.728 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:26:36.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:35 vm08.local ceph-mon[50506]: pgmap v1974: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:36.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:35 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:26:36.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:35 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:26:36.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:35 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:26:36.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:35 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:26:36.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:35 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:26:36.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:35 vm01.local ceph-mon[48597]: pgmap v1974: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:36.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:35 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:26:36.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:35 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:26:36.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:35 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:26:36.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:35 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:26:36.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:35 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:26:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:37 vm08.local ceph-mon[50506]: pgmap v1975: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:26:38.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:37 vm01.local ceph-mon[48597]: pgmap v1975: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:26:39.729 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:26:39.730 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:26:39.756 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:26:39.757 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:26:40.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:39 vm08.local ceph-mon[50506]: pgmap v1976: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:40.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:39 vm01.local ceph-mon[48597]: pgmap v1976: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:41 vm08.local ceph-mon[50506]: pgmap v1977: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:26:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:41 vm01.local ceph-mon[48597]: pgmap v1977: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:26:44.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:43 vm01.local ceph-mon[48597]: pgmap v1978: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:44.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:43 vm08.local ceph-mon[50506]: pgmap v1978: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:44.758 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:26:44.759 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:26:44.786 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:26:44.786 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:26:45.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:44 vm08.local ceph-mon[50506]: pgmap v1979: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:45.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:44 vm01.local ceph-mon[48597]: pgmap v1979: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:47 vm08.local ceph-mon[50506]: pgmap v1980: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:26:48.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:47 vm01.local ceph-mon[48597]: pgmap v1980: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:26:49.788 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:26:49.789 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:26:49.818 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:26:49.819 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:26:50.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:49 vm08.local ceph-mon[50506]: pgmap v1981: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:50.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:49 vm01.local ceph-mon[48597]: pgmap v1981: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:52.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:51 vm08.local ceph-mon[50506]: pgmap v1982: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:26:52.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:51 vm01.local ceph-mon[48597]: pgmap v1982: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:26:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:53 vm01.local ceph-mon[48597]: pgmap v1983: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:54.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:53 vm08.local ceph-mon[50506]: pgmap v1983: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:54.821 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:26:54.821 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:26:54.848 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:26:54.849 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:26:56.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:55 vm08.local ceph-mon[50506]: pgmap v1984: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:56.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:55 vm01.local ceph-mon[48597]: pgmap v1984: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:57 vm08.local ceph-mon[50506]: pgmap v1985: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:26:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:57 vm01.local ceph-mon[48597]: pgmap v1985: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:26:59.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:26:59 vm08.local ceph-mon[50506]: pgmap v1986: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:59.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:26:59 vm01.local ceph-mon[48597]: pgmap v1986: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:26:59.850 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:26:59.851 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:26:59.878 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:26:59.879 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:27:02.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:01 vm08.local ceph-mon[50506]: pgmap v1987: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:27:02.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:01 vm01.local ceph-mon[48597]: pgmap v1987: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:27:04.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:03 vm01.local ceph-mon[48597]: pgmap v1988: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:04.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:03 vm08.local ceph-mon[50506]: pgmap v1988: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:04.880 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:27:04.881 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:27:04.908 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:27:04.909 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:27:06.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:05 vm08.local ceph-mon[50506]: pgmap v1989: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:06.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:05 vm01.local ceph-mon[48597]: pgmap v1989: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:08.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:07 vm08.local ceph-mon[50506]: pgmap v1990: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:27:08.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:07 vm01.local ceph-mon[48597]: pgmap v1990: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:27:09.910 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:27:09.911 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:27:09.949 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:27:09.950 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:27:10.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:09 vm08.local ceph-mon[50506]: pgmap v1991: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:10.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:09 vm01.local ceph-mon[48597]: pgmap v1991: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:27:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:27:11.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:27:11.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:27:12.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:12 vm08.local ceph-mon[50506]: pgmap v1992: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:27:12.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:12 vm01.local ceph-mon[48597]: pgmap v1992: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:27:14.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:14 vm08.local ceph-mon[50506]: pgmap v1993: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:14.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:14 vm01.local ceph-mon[48597]: pgmap v1993: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:14.952 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:27:14.952 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:27:14.980 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:27:14.981 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:27:15.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:15 vm08.local ceph-mon[50506]: pgmap v1994: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:15.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:15 vm01.local ceph-mon[48597]: pgmap v1994: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:18.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:18 vm08.local ceph-mon[50506]: pgmap v1995: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:27:18.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:18 vm01.local ceph-mon[48597]: pgmap v1995: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:27:19.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:19 vm08.local ceph-mon[50506]: pgmap v1996: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:19 vm01.local ceph-mon[48597]: pgmap v1996: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:19.983 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:27:19.983 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:27:20.010 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:27:20.010 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:27:22.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:21 vm08.local ceph-mon[50506]: pgmap v1997: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:27:22.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:21 vm01.local ceph-mon[48597]: pgmap v1997: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:27:24.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:24 vm01.local ceph-mon[48597]: pgmap v1998: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:24.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:24 vm08.local ceph-mon[50506]: pgmap v1998: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:25.011 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:27:25.012 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:27:25.038 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:27:25.038 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:27:25.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:25 vm08.local ceph-mon[50506]: pgmap v1999: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:25.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:25 vm01.local ceph-mon[48597]: pgmap v1999: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:28.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:27 vm08.local ceph-mon[50506]: pgmap v2000: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:27:28.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:27 vm01.local ceph-mon[48597]: pgmap v2000: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:27:30.040 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:27:30.040 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:27:30.137 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:27:30.138 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:27:30.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:30 vm08.local ceph-mon[50506]: pgmap v2001: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:30.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:30 vm01.local ceph-mon[48597]: pgmap v2001: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:31.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:31 vm08.local ceph-mon[50506]: pgmap v2002: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:27:31.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:31 vm01.local ceph-mon[48597]: pgmap v2002: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:27:34.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:33 vm01.local ceph-mon[48597]: pgmap v2003: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:34.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:33 vm08.local ceph-mon[50506]: pgmap v2003: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:35.139 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:27:35.140 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:27:35.170 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:27:35.170 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:27:36.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:35 vm08.local ceph-mon[50506]: pgmap v2004: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:36.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:35 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:27:36.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:35 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:27:36.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:35 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:27:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:35 vm01.local ceph-mon[48597]: pgmap v2004: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:35 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:27:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:35 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:27:36.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:35 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:27:37.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:36 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:27:37.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:36 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:27:37.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:36 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:27:37.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:36 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:27:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:37 vm08.local ceph-mon[50506]: pgmap v2005: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:27:38.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:37 vm01.local ceph-mon[48597]: pgmap v2005: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:27:40.172 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:27:40.173 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:27:40.200 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:27:40.201 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:27:40.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:39 vm08.local ceph-mon[50506]: pgmap v2006: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:40.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:39 vm01.local ceph-mon[48597]: pgmap v2006: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:41.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:41 vm08.local ceph-mon[50506]: pgmap v2007: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:27:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:41 vm01.local ceph-mon[48597]: pgmap v2007: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:27:44.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:43 vm01.local ceph-mon[48597]: pgmap v2008: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:44.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:43 vm08.local ceph-mon[50506]: pgmap v2008: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:45.202 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:27:45.203 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:27:45.229 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:27:45.230 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:27:46.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:45 vm08.local ceph-mon[50506]: pgmap v2009: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:46.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:45 vm01.local ceph-mon[48597]: pgmap v2009: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:47.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:47 vm08.local ceph-mon[50506]: pgmap v2010: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:27:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:47 vm01.local ceph-mon[48597]: pgmap v2010: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:27:50.232 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:27:50.232 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:27:50.259 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:27:50.259 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:27:50.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:49 vm08.local ceph-mon[50506]: pgmap v2011: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:50.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:49 vm01.local ceph-mon[48597]: pgmap v2011: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:52.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:51 vm08.local ceph-mon[50506]: pgmap v2012: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:27:52.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:51 vm01.local ceph-mon[48597]: pgmap v2012: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:27:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:53 vm01.local ceph-mon[48597]: pgmap v2013: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:54.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:53 vm08.local ceph-mon[50506]: pgmap v2013: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:55.261 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:27:55.261 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:27:55.286 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:27:55.287 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:27:55.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:55 vm08.local ceph-mon[50506]: pgmap v2014: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:55 vm01.local ceph-mon[48597]: pgmap v2014: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:57 vm08.local ceph-mon[50506]: pgmap v2015: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:27:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:57 vm01.local ceph-mon[48597]: pgmap v2015: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:27:59.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:27:59 vm01.local ceph-mon[48597]: pgmap v2016: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:27:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:27:59 vm08.local ceph-mon[50506]: pgmap v2016: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:00.288 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:28:00.289 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:28:00.314 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:28:00.315 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:28:02.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:01 vm08.local ceph-mon[50506]: pgmap v2017: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:28:02.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:01 vm01.local ceph-mon[48597]: pgmap v2017: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:28:04.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:03 vm01.local ceph-mon[48597]: pgmap v2018: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:04.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:03 vm08.local ceph-mon[50506]: pgmap v2018: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:05.316 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:28:05.332 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:28:05.380 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:28:05.380 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:28:06.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:05 vm08.local ceph-mon[50506]: pgmap v2019: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:06.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:05 vm01.local ceph-mon[48597]: pgmap v2019: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:08.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:07 vm08.local ceph-mon[50506]: pgmap v2020: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:28:08.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:07 vm01.local ceph-mon[48597]: pgmap v2020: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:28:10.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:09 vm08.local ceph-mon[50506]: pgmap v2021: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:10.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:09 vm01.local ceph-mon[48597]: pgmap v2021: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:10.382 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:28:10.382 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:28:10.408 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:28:10.408 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:28:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:28:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:28:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:28:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:28:12.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:11 vm08.local ceph-mon[50506]: pgmap v2022: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:28:12.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:11 vm01.local ceph-mon[48597]: pgmap v2022: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:28:14.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:13 vm01.local ceph-mon[48597]: pgmap v2023: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:14.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:13 vm08.local ceph-mon[50506]: pgmap v2023: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:15.410 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:28:15.410 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:28:15.439 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:28:15.440 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:28:16.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:16 vm08.local ceph-mon[50506]: pgmap v2024: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:16.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:16 vm01.local ceph-mon[48597]: pgmap v2024: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:17 vm08.local ceph-mon[50506]: pgmap v2025: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:28:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:17 vm01.local ceph-mon[48597]: pgmap v2025: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:28:19.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:19 vm08.local ceph-mon[50506]: pgmap v2026: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:19.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:19 vm01.local ceph-mon[48597]: pgmap v2026: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:20.441 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:28:20.442 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:28:20.469 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:28:20.469 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:28:22.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:21 vm08.local ceph-mon[50506]: pgmap v2027: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:28:22.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:21 vm01.local ceph-mon[48597]: pgmap v2027: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:28:24.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:23 vm01.local ceph-mon[48597]: pgmap v2028: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:24.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:23 vm08.local ceph-mon[50506]: pgmap v2028: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:25.471 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:28:25.471 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:28:25.498 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:28:25.498 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:28:25.942 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:25 vm01.local ceph-mon[48597]: pgmap v2029: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:26.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:25 vm08.local ceph-mon[50506]: pgmap v2029: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:28.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:27 vm08.local ceph-mon[50506]: pgmap v2030: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:28:28.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:27 vm01.local ceph-mon[48597]: pgmap v2030: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:28:29.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:29 vm08.local ceph-mon[50506]: pgmap v2031: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:29.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:29 vm01.local ceph-mon[48597]: pgmap v2031: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:30.500 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:28:30.500 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:28:30.525 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:28:30.526 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:28:32.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:32 vm08.local ceph-mon[50506]: pgmap v2032: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:32.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:32 vm01.local ceph-mon[48597]: pgmap v2032: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:33 vm08.local ceph-mon[50506]: pgmap v2033: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:28:33.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:33 vm01.local ceph-mon[48597]: pgmap v2033: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:28:35.527 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:28:35.528 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:28:35.555 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:28:35.556 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:28:36.199 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:35 vm08.local ceph-mon[50506]: pgmap v2034: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:28:36.203 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:35 vm01.local ceph-mon[48597]: pgmap v2034: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:28:37.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:36 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:28:37.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:36 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:28:37.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:36 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:28:37.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:36 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:28:37.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:36 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:28:37.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:36 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:28:37.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:36 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:28:37.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:36 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:28:37.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:36 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:28:37.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:36 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:28:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:37 vm08.local ceph-mon[50506]: pgmap v2035: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:38.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:37 vm01.local ceph-mon[48597]: pgmap v2035: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:39.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:39 vm08.local ceph-mon[50506]: pgmap v2036: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:28:39.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:39 vm01.local ceph-mon[48597]: pgmap v2036: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:28:40.557 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:28:40.557 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:28:40.587 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:28:40.587 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:28:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:41 vm08.local ceph-mon[50506]: pgmap v2037: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:41 vm01.local ceph-mon[48597]: pgmap v2037: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:44.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:43 vm01.local ceph-mon[48597]: pgmap v2038: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:44.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:43 vm08.local ceph-mon[50506]: pgmap v2038: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:45.588 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:28:45.589 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:28:45.617 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:28:45.617 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:28:46.146 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:45 vm08.local ceph-mon[50506]: pgmap v2039: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:46.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:45 vm01.local ceph-mon[48597]: pgmap v2039: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:47 vm08.local ceph-mon[50506]: pgmap v2040: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:28:48.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:47 vm01.local ceph-mon[48597]: pgmap v2040: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:28:49.269 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:49 vm01.local ceph-mon[48597]: pgmap v2041: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:49.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:49 vm08.local ceph-mon[50506]: pgmap v2041: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:50.618 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:28:50.619 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:28:50.645 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:28:50.646 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:28:52.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:52 vm08.local ceph-mon[50506]: pgmap v2042: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:28:52.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:52 vm01.local ceph-mon[48597]: pgmap v2042: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:28:53.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:53 vm08.local ceph-mon[50506]: pgmap v2043: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:53.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:53 vm01.local ceph-mon[48597]: pgmap v2043: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:55.647 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:28:55.647 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:28:55.673 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:28:55.673 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:28:56.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:55 vm08.local ceph-mon[50506]: pgmap v2044: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:56.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:55 vm01.local ceph-mon[48597]: pgmap v2044: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:57 vm08.local ceph-mon[50506]: pgmap v2045: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:28:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:57 vm01.local ceph-mon[48597]: pgmap v2045: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:28:59.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:28:59 vm01.local ceph-mon[48597]: pgmap v2046: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:28:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:28:59 vm08.local ceph-mon[50506]: pgmap v2046: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:00.674 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:29:00.675 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:29:00.702 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:29:00.703 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:29:02.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:01 vm08.local ceph-mon[50506]: pgmap v2047: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:29:02.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:01 vm01.local ceph-mon[48597]: pgmap v2047: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:29:04.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:03 vm01.local ceph-mon[48597]: pgmap v2048: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:04.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:03 vm08.local ceph-mon[50506]: pgmap v2048: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:05.704 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:29:05.705 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:29:05.731 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:29:05.732 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:29:06.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:05 vm08.local ceph-mon[50506]: pgmap v2049: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:06.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:05 vm01.local ceph-mon[48597]: pgmap v2049: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:08.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:07 vm08.local ceph-mon[50506]: pgmap v2050: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:29:08.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:07 vm01.local ceph-mon[48597]: pgmap v2050: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:29:09.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:09 vm01.local ceph-mon[48597]: pgmap v2051: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:09 vm08.local ceph-mon[50506]: pgmap v2051: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:10.733 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:29:10.734 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:29:10.759 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:29:10.759 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:29:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:29:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:29:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:29:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:11 vm08.local ceph-mon[50506]: pgmap v2052: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:29:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:29:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:11 vm01.local ceph-mon[48597]: pgmap v2052: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:29:14.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:13 vm01.local ceph-mon[48597]: pgmap v2053: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:14.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:13 vm08.local ceph-mon[50506]: pgmap v2053: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:15.760 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:29:15.761 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:29:15.787 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:29:15.788 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:29:16.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:15 vm08.local ceph-mon[50506]: pgmap v2054: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:16.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:15 vm01.local ceph-mon[48597]: pgmap v2054: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:18.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:17 vm08.local ceph-mon[50506]: pgmap v2055: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:29:18.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:17 vm01.local ceph-mon[48597]: pgmap v2055: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:29:19.269 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:19 vm01.local ceph-mon[48597]: pgmap v2056: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:19.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:19 vm08.local ceph-mon[50506]: pgmap v2056: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:20.789 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:29:20.789 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:29:20.814 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:29:20.814 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:29:22.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:21 vm08.local ceph-mon[50506]: pgmap v2057: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:29:22.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:21 vm01.local ceph-mon[48597]: pgmap v2057: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:29:24.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:23 vm01.local ceph-mon[48597]: pgmap v2058: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:24.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:23 vm08.local ceph-mon[50506]: pgmap v2058: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:25.816 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:29:25.816 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:29:25.843 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:29:25.844 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:29:26.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:25 vm08.local ceph-mon[50506]: pgmap v2059: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:26.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:25 vm01.local ceph-mon[48597]: pgmap v2059: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:28.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:27 vm08.local ceph-mon[50506]: pgmap v2060: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:29:28.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:27 vm01.local ceph-mon[48597]: pgmap v2060: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:29:29.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:29 vm01.local ceph-mon[48597]: pgmap v2061: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:29 vm08.local ceph-mon[50506]: pgmap v2061: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:30.845 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:29:30.846 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:29:30.873 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:29:30.873 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:29:32.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:31 vm08.local ceph-mon[50506]: pgmap v2062: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:29:32.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:31 vm01.local ceph-mon[48597]: pgmap v2062: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:29:34.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:33 vm01.local ceph-mon[48597]: pgmap v2063: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:34.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:33 vm08.local ceph-mon[50506]: pgmap v2063: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:35.875 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:29:35.882 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:29:36.051 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:29:36.052 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:29:36.449 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:36 vm08.local ceph-mon[50506]: pgmap v2064: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:36.451 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:36 vm01.local ceph-mon[48597]: pgmap v2064: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:37 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:29:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:37 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:29:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:37 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:29:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:37 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:29:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:37 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:29:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:37 vm08.local ceph-mon[50506]: pgmap v2065: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:29:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:37 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:29:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:37 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:29:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:37 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:29:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:37 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:29:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:37 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:29:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:37 vm01.local ceph-mon[48597]: pgmap v2065: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:29:39.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:39 vm01.local ceph-mon[48597]: pgmap v2066: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:39 vm08.local ceph-mon[50506]: pgmap v2066: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:41.053 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:29:41.053 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:29:41.080 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:29:41.080 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:29:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:41 vm08.local ceph-mon[50506]: pgmap v2067: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:29:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:41 vm01.local ceph-mon[48597]: pgmap v2067: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:29:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:43 vm08.local ceph-mon[50506]: pgmap v2068: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:43.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:43 vm01.local ceph-mon[48597]: pgmap v2068: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:46.082 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:29:46.082 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:29:46.108 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:29:46.109 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:29:46.297 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:45 vm08.local ceph-mon[50506]: pgmap v2069: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:46.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:45 vm01.local ceph-mon[48597]: pgmap v2069: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:47 vm08.local ceph-mon[50506]: pgmap v2070: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:47 vm01.local ceph-mon[48597]: pgmap v2070: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:29:49.269 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:49 vm01.local ceph-mon[48597]: pgmap v2071: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:49.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:49 vm08.local ceph-mon[50506]: pgmap v2071: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:51.110 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:29:51.111 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:29:51.137 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:29:51.138 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:29:52.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:51 vm08.local ceph-mon[50506]: pgmap v2072: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:29:52.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:51 vm01.local ceph-mon[48597]: pgmap v2072: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:29:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:53 vm01.local ceph-mon[48597]: pgmap v2073: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:54.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:53 vm08.local ceph-mon[50506]: pgmap v2073: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:56.139 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:29:56.139 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:29:56.165 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:29:56.166 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:29:56.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:55 vm08.local ceph-mon[50506]: pgmap v2074: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:56.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:55 vm01.local ceph-mon[48597]: pgmap v2074: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:57 vm08.local ceph-mon[50506]: pgmap v2075: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:29:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:57 vm01.local ceph-mon[48597]: pgmap v2075: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:29:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:29:59 vm08.local ceph-mon[50506]: pgmap v2076: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:29:59.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:29:59 vm01.local ceph-mon[48597]: pgmap v2076: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:00.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:00 vm08.local ceph-mon[50506]: overall HEALTH_OK 2026-03-10T11:30:00.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:00 vm01.local ceph-mon[48597]: overall HEALTH_OK 2026-03-10T11:30:01.167 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:30:01.168 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:30:01.195 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:30:01.196 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:30:01.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:01 vm08.local ceph-mon[50506]: pgmap v2077: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:30:01.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:01 vm01.local ceph-mon[48597]: pgmap v2077: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:30:04.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:03 vm01.local ceph-mon[48597]: pgmap v2078: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:04.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:03 vm08.local ceph-mon[50506]: pgmap v2078: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:06.197 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:30:06.198 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:30:06.225 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:30:06.226 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:30:06.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:05 vm08.local ceph-mon[50506]: pgmap v2079: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:06.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:05 vm01.local ceph-mon[48597]: pgmap v2079: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:08.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:07 vm08.local ceph-mon[50506]: pgmap v2080: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:30:08.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:07 vm01.local ceph-mon[48597]: pgmap v2080: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:30:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:09 vm08.local ceph-mon[50506]: pgmap v2081: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:09.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:09 vm01.local ceph-mon[48597]: pgmap v2081: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:10.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:30:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:30:11.228 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:30:11.228 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:30:11.255 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:30:11.255 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:30:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:30:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:11 vm08.local ceph-mon[50506]: pgmap v2082: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:30:11.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:30:11.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:11 vm01.local ceph-mon[48597]: pgmap v2082: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:30:14.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:13 vm01.local ceph-mon[48597]: pgmap v2083: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:14.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:13 vm08.local ceph-mon[50506]: pgmap v2083: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:16.257 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:30:16.257 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:30:16.283 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:30:16.283 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:30:16.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:15 vm08.local ceph-mon[50506]: pgmap v2084: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:16.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:15 vm01.local ceph-mon[48597]: pgmap v2084: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:18.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:17 vm08.local ceph-mon[50506]: pgmap v2085: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:30:18.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:17 vm01.local ceph-mon[48597]: pgmap v2085: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:30:19.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:19 vm08.local ceph-mon[50506]: pgmap v2086: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:19 vm01.local ceph-mon[48597]: pgmap v2086: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:21.285 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:30:21.285 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:30:21.313 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:30:21.314 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:30:22.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:21 vm08.local ceph-mon[50506]: pgmap v2087: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:30:22.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:21 vm01.local ceph-mon[48597]: pgmap v2087: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:30:24.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:23 vm01.local ceph-mon[48597]: pgmap v2088: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:24.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:23 vm08.local ceph-mon[50506]: pgmap v2088: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:26.315 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:30:26.316 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:30:26.341 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:30:26.341 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:30:26.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:25 vm08.local ceph-mon[50506]: pgmap v2089: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:26.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:25 vm01.local ceph-mon[48597]: pgmap v2089: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:28.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:27 vm08.local ceph-mon[50506]: pgmap v2090: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:30:28.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:27 vm01.local ceph-mon[48597]: pgmap v2090: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:30:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:29 vm08.local ceph-mon[50506]: pgmap v2091: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:29.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:29 vm01.local ceph-mon[48597]: pgmap v2091: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:31.343 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:30:31.343 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:30:31.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:31 vm01.local ceph-mon[48597]: pgmap v2092: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:30:31.370 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:30:31.371 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:30:31.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:31 vm08.local ceph-mon[50506]: pgmap v2092: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:30:34.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:33 vm01.local ceph-mon[48597]: pgmap v2093: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:34.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:33 vm08.local ceph-mon[50506]: pgmap v2093: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:36.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:35 vm08.local ceph-mon[50506]: pgmap v2094: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:36.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:35 vm01.local ceph-mon[48597]: pgmap v2094: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:36.372 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:30:36.372 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:30:36.406 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:30:36.406 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:30:37.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:36 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:37.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:36 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:37.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:36 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:37.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:36 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:37.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:36 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:37.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:36 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:37 vm08.local ceph-mon[50506]: pgmap v2095: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:30:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:37 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:30:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:37 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:30:38.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:37 vm01.local ceph-mon[48597]: pgmap v2095: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:30:38.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:37 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:30:38.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:37 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:30:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:39 vm08.local ceph-mon[50506]: pgmap v2096: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:39.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:39 vm01.local ceph-mon[48597]: pgmap v2096: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:41.408 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:30:41.408 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:30:41.434 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:30:41.435 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:30:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:41 vm08.local ceph-mon[50506]: pgmap v2097: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:30:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:41 vm01.local ceph-mon[48597]: pgmap v2097: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:30:44.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:43 vm01.local ceph-mon[48597]: pgmap v2098: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:44.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:43 vm08.local ceph-mon[50506]: pgmap v2098: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:46.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:45 vm08.local ceph-mon[50506]: pgmap v2099: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:46.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:45 vm01.local ceph-mon[48597]: pgmap v2099: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:46.436 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:30:46.436 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:30:46.463 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:30:46.464 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:30:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:47 vm08.local ceph-mon[50506]: pgmap v2100: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:30:48.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:47 vm01.local ceph-mon[48597]: pgmap v2100: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:30:49.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:49 vm08.local ceph-mon[50506]: pgmap v2101: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:49.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:49 vm01.local ceph-mon[48597]: pgmap v2101: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:51.465 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:30:51.465 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:30:51.492 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:30:51.493 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:30:52.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:51 vm08.local ceph-mon[50506]: pgmap v2102: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:30:52.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:51 vm01.local ceph-mon[48597]: pgmap v2102: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:30:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:53 vm01.local ceph-mon[48597]: pgmap v2103: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:54.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:53 vm08.local ceph-mon[50506]: pgmap v2103: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:56.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:55 vm08.local ceph-mon[50506]: pgmap v2104: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:56.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:55 vm01.local ceph-mon[48597]: pgmap v2104: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:56.494 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:30:56.495 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:30:56.521 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:30:56.522 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:30:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:57 vm08.local ceph-mon[50506]: pgmap v2105: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:30:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:57 vm01.local ceph-mon[48597]: pgmap v2105: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:30:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:30:59 vm08.local ceph-mon[50506]: pgmap v2106: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:30:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:30:59 vm01.local ceph-mon[48597]: pgmap v2106: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:01.523 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:31:01.523 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:31:01.550 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:31:01.551 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:31:02.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:01 vm08.local ceph-mon[50506]: pgmap v2107: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:31:02.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:01 vm01.local ceph-mon[48597]: pgmap v2107: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:31:04.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:03 vm01.local ceph-mon[48597]: pgmap v2108: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:04.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:03 vm08.local ceph-mon[50506]: pgmap v2108: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:06.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:05 vm08.local ceph-mon[50506]: pgmap v2109: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:06.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:05 vm01.local ceph-mon[48597]: pgmap v2109: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:06.552 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:31:06.552 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:31:06.581 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:31:06.582 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:31:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:07 vm08.local ceph-mon[50506]: pgmap v2110: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:31:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:07 vm01.local ceph-mon[48597]: pgmap v2110: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:31:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:09 vm08.local ceph-mon[50506]: pgmap v2111: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:09.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:09 vm01.local ceph-mon[48597]: pgmap v2111: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:31:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:31:11.583 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:31:11.585 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:31:11.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:31:11.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:11 vm01.local ceph-mon[48597]: pgmap v2112: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:31:11.665 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:31:11.666 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:31:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:31:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:11 vm08.local ceph-mon[50506]: pgmap v2112: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:31:14.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:13 vm01.local ceph-mon[48597]: pgmap v2113: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:14.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:13 vm08.local ceph-mon[50506]: pgmap v2113: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:16.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:15 vm08.local ceph-mon[50506]: pgmap v2114: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:16.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:15 vm01.local ceph-mon[48597]: pgmap v2114: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:16.667 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:31:16.668 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:31:16.700 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:31:16.701 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:31:18.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:17 vm08.local ceph-mon[50506]: pgmap v2115: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:31:18.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:17 vm01.local ceph-mon[48597]: pgmap v2115: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:31:19.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:19 vm08.local ceph-mon[50506]: pgmap v2116: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:19 vm01.local ceph-mon[48597]: pgmap v2116: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:21.702 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:31:21.703 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:31:21.729 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:31:21.729 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:31:22.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:21 vm08.local ceph-mon[50506]: pgmap v2117: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:31:22.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:21 vm01.local ceph-mon[48597]: pgmap v2117: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:31:24.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:24 vm01.local ceph-mon[48597]: pgmap v2118: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:24.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:24 vm08.local ceph-mon[50506]: pgmap v2118: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:26.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:26 vm08.local ceph-mon[50506]: pgmap v2119: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:26.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:26 vm01.local ceph-mon[48597]: pgmap v2119: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:26.730 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:31:26.731 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:31:26.762 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:31:26.762 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:31:27.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:27 vm08.local ceph-mon[50506]: pgmap v2120: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:31:27.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:27 vm01.local ceph-mon[48597]: pgmap v2120: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:31:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:29 vm08.local ceph-mon[50506]: pgmap v2121: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:29.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:29 vm01.local ceph-mon[48597]: pgmap v2121: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:31.763 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:31:31.764 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:31:31.790 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:31:31.791 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:31:32.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:31 vm08.local ceph-mon[50506]: pgmap v2122: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:31:32.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:31 vm01.local ceph-mon[48597]: pgmap v2122: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:31:34.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:33 vm01.local ceph-mon[48597]: pgmap v2123: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:34.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:33 vm08.local ceph-mon[50506]: pgmap v2123: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:36.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:35 vm08.local ceph-mon[50506]: pgmap v2124: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:36.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:35 vm01.local ceph-mon[48597]: pgmap v2124: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:36.792 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:31:36.792 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:31:36.819 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:31:36.820 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:31:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:37 vm08.local ceph-mon[50506]: pgmap v2125: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:31:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:37 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:37 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:37 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:37 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:31:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:37 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:31:38.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:37 vm01.local ceph-mon[48597]: pgmap v2125: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:31:38.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:37 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:38.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:37 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:38.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:37 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:38.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:37 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:31:38.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:37 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:31:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:39 vm08.local ceph-mon[50506]: pgmap v2126: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:39.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:39 vm01.local ceph-mon[48597]: pgmap v2126: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:41.821 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:31:41.821 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:31:41.847 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:31:41.848 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:31:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:41 vm08.local ceph-mon[50506]: pgmap v2127: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:31:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:41 vm01.local ceph-mon[48597]: pgmap v2127: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:31:44.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:44 vm01.local ceph-mon[48597]: pgmap v2128: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:44.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:44 vm08.local ceph-mon[50506]: pgmap v2128: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:46.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:46 vm08.local ceph-mon[50506]: pgmap v2129: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:46.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:46 vm01.local ceph-mon[48597]: pgmap v2129: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:46.849 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:31:46.850 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:31:46.877 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:31:46.877 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:31:47.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:47 vm08.local ceph-mon[50506]: pgmap v2130: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:31:47.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:47 vm01.local ceph-mon[48597]: pgmap v2130: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:31:49.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:49 vm08.local ceph-mon[50506]: pgmap v2131: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:49.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:49 vm01.local ceph-mon[48597]: pgmap v2131: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:51.879 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:31:51.880 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:31:51.906 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:31:51.906 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:31:52.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:51 vm08.local ceph-mon[50506]: pgmap v2132: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:31:52.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:51 vm01.local ceph-mon[48597]: pgmap v2132: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:31:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:53 vm01.local ceph-mon[48597]: pgmap v2133: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:54.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:53 vm08.local ceph-mon[50506]: pgmap v2133: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:56.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:55 vm08.local ceph-mon[50506]: pgmap v2134: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:56.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:55 vm01.local ceph-mon[48597]: pgmap v2134: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:56.908 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:31:56.909 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:31:56.949 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:31:56.950 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:31:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:57 vm08.local ceph-mon[50506]: pgmap v2135: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:31:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:57 vm01.local ceph-mon[48597]: pgmap v2135: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:31:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:31:59 vm08.local ceph-mon[50506]: pgmap v2136: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:31:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:31:59 vm01.local ceph-mon[48597]: pgmap v2136: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:01.951 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:32:01.952 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:32:01.978 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:32:01.979 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:32:02.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:01 vm08.local ceph-mon[50506]: pgmap v2137: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:32:02.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:01 vm01.local ceph-mon[48597]: pgmap v2137: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:32:04.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:03 vm01.local ceph-mon[48597]: pgmap v2138: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:04.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:03 vm08.local ceph-mon[50506]: pgmap v2138: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:06.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:05 vm08.local ceph-mon[50506]: pgmap v2139: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:06.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:05 vm01.local ceph-mon[48597]: pgmap v2139: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:06.980 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:32:06.981 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:32:07.007 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:32:07.007 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:32:08.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:07 vm08.local ceph-mon[50506]: pgmap v2140: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:32:08.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:07 vm01.local ceph-mon[48597]: pgmap v2140: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:32:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:09 vm08.local ceph-mon[50506]: pgmap v2141: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:09.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:09 vm01.local ceph-mon[48597]: pgmap v2141: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:10.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:32:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:32:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:32:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:11 vm08.local ceph-mon[50506]: pgmap v2142: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:32:11.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:32:11.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:11 vm01.local ceph-mon[48597]: pgmap v2142: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:32:12.009 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:32:12.009 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:32:12.036 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:32:12.037 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:32:14.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:13 vm01.local ceph-mon[48597]: pgmap v2143: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:14.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:13 vm08.local ceph-mon[50506]: pgmap v2143: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:16.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:15 vm08.local ceph-mon[50506]: pgmap v2144: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:16.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:15 vm01.local ceph-mon[48597]: pgmap v2144: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:17.038 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:32:17.039 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:32:17.066 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:32:17.066 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:32:18.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:17 vm08.local ceph-mon[50506]: pgmap v2145: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:32:18.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:17 vm01.local ceph-mon[48597]: pgmap v2145: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:32:19.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:19 vm08.local ceph-mon[50506]: pgmap v2146: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:19 vm01.local ceph-mon[48597]: pgmap v2146: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:22.067 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:32:22.068 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:32:22.093 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:32:22.093 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:32:22.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:21 vm08.local ceph-mon[50506]: pgmap v2147: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:32:22.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:21 vm01.local ceph-mon[48597]: pgmap v2147: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:32:24.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:23 vm01.local ceph-mon[48597]: pgmap v2148: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:24.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:23 vm08.local ceph-mon[50506]: pgmap v2148: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:26.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:25 vm08.local ceph-mon[50506]: pgmap v2149: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:26.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:25 vm01.local ceph-mon[48597]: pgmap v2149: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:27.095 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:32:27.096 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:32:27.171 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:32:27.171 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:32:28.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:27 vm08.local ceph-mon[50506]: pgmap v2150: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:32:28.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:27 vm01.local ceph-mon[48597]: pgmap v2150: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:32:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:29 vm08.local ceph-mon[50506]: pgmap v2151: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:29.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:29 vm01.local ceph-mon[48597]: pgmap v2151: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:32.173 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:32:32.173 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:32:32.202 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:32:32.203 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:32:32.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:31 vm08.local ceph-mon[50506]: pgmap v2152: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:32:32.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:31 vm01.local ceph-mon[48597]: pgmap v2152: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:32:34.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:33 vm01.local ceph-mon[48597]: pgmap v2153: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:34.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:33 vm08.local ceph-mon[50506]: pgmap v2153: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:36.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:35 vm08.local ceph-mon[50506]: pgmap v2154: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:36.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:35 vm01.local ceph-mon[48597]: pgmap v2154: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:37.204 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:32:37.205 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:32:37.232 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:32:37.232 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:32:38.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:38 vm01.local ceph-mon[48597]: pgmap v2155: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:32:38.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:38 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:32:38.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:38 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:32:38.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:38 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:32:38.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:38 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:32:38.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:38 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:32:38.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:38 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:32:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:38 vm08.local ceph-mon[50506]: pgmap v2155: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:32:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:38 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:32:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:38 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:32:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:38 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:32:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:38 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:32:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:38 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:32:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:38 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:32:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:39 vm08.local ceph-mon[50506]: pgmap v2156: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:39.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:39 vm01.local ceph-mon[48597]: pgmap v2156: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:42.233 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:32:42.234 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:32:42.268 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:32:42.269 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:32:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:41 vm08.local ceph-mon[50506]: pgmap v2157: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:32:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:41 vm01.local ceph-mon[48597]: pgmap v2157: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:32:44.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:43 vm01.local ceph-mon[48597]: pgmap v2158: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:44.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:43 vm08.local ceph-mon[50506]: pgmap v2158: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:46.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:46 vm08.local ceph-mon[50506]: pgmap v2159: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:46.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:46 vm01.local ceph-mon[48597]: pgmap v2159: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:47.270 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:32:47.270 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:32:47.296 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:32:47.296 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:32:47.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:47 vm08.local ceph-mon[50506]: pgmap v2160: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:32:47.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:47 vm01.local ceph-mon[48597]: pgmap v2160: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:32:49.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:49 vm08.local ceph-mon[50506]: pgmap v2161: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:49.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:49 vm01.local ceph-mon[48597]: pgmap v2161: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:52.298 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:32:52.299 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:32:52.325 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:32:52.325 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:32:52.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:51 vm08.local ceph-mon[50506]: pgmap v2162: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:32:52.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:51 vm01.local ceph-mon[48597]: pgmap v2162: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:32:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:53 vm01.local ceph-mon[48597]: pgmap v2163: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:54.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:53 vm08.local ceph-mon[50506]: pgmap v2163: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:56.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:55 vm08.local ceph-mon[50506]: pgmap v2164: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:56.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:55 vm01.local ceph-mon[48597]: pgmap v2164: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:57.327 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:32:57.327 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:32:57.354 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:32:57.355 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:32:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:57 vm08.local ceph-mon[50506]: pgmap v2165: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:32:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:57 vm01.local ceph-mon[48597]: pgmap v2165: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:32:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:32:59 vm08.local ceph-mon[50506]: pgmap v2166: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:32:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:32:59 vm01.local ceph-mon[48597]: pgmap v2166: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:02.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:01 vm08.local ceph-mon[50506]: pgmap v2167: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:33:02.356 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:33:02.357 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:33:02.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:01 vm01.local ceph-mon[48597]: pgmap v2167: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:33:02.385 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:33:02.386 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:33:04.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:04 vm01.local ceph-mon[48597]: pgmap v2168: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:04.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:04 vm08.local ceph-mon[50506]: pgmap v2168: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:06.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:06 vm08.local ceph-mon[50506]: pgmap v2169: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:06.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:06 vm01.local ceph-mon[48597]: pgmap v2169: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:07.387 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:33:07.388 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:33:07.415 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:33:07.415 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:33:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:07 vm08.local ceph-mon[50506]: pgmap v2170: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:33:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:07 vm01.local ceph-mon[48597]: pgmap v2170: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:33:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:09 vm08.local ceph-mon[50506]: pgmap v2171: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:09.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:09 vm01.local ceph-mon[48597]: pgmap v2171: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:10.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:33:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:33:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:33:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:11 vm08.local ceph-mon[50506]: pgmap v2172: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:33:11.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:33:11.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:11 vm01.local ceph-mon[48597]: pgmap v2172: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:33:12.417 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:33:12.417 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:33:12.446 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:33:12.446 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:33:14.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:13 vm01.local ceph-mon[48597]: pgmap v2173: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:14.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:13 vm08.local ceph-mon[50506]: pgmap v2173: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:16.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:15 vm08.local ceph-mon[50506]: pgmap v2174: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:16.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:15 vm01.local ceph-mon[48597]: pgmap v2174: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:17.448 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:33:17.449 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:33:17.477 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:33:17.478 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:33:18.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:17 vm08.local ceph-mon[50506]: pgmap v2175: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:33:18.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:17 vm01.local ceph-mon[48597]: pgmap v2175: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:33:19.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:19 vm08.local ceph-mon[50506]: pgmap v2176: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:19 vm01.local ceph-mon[48597]: pgmap v2176: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:22.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:21 vm08.local ceph-mon[50506]: pgmap v2177: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:33:22.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:21 vm01.local ceph-mon[48597]: pgmap v2177: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:33:22.479 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:33:22.480 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:33:22.508 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:33:22.509 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:33:24.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:23 vm01.local ceph-mon[48597]: pgmap v2178: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:24.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:23 vm08.local ceph-mon[50506]: pgmap v2178: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:26.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:25 vm08.local ceph-mon[50506]: pgmap v2179: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:26.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:25 vm01.local ceph-mon[48597]: pgmap v2179: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:27.510 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:33:27.511 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:33:27.536 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:33:27.536 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:33:28.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:27 vm08.local ceph-mon[50506]: pgmap v2180: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:33:28.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:27 vm01.local ceph-mon[48597]: pgmap v2180: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:33:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:29 vm08.local ceph-mon[50506]: pgmap v2181: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:29.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:29 vm01.local ceph-mon[48597]: pgmap v2181: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:32.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:31 vm08.local ceph-mon[50506]: pgmap v2182: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:33:32.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:31 vm01.local ceph-mon[48597]: pgmap v2182: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:33:32.538 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:33:32.538 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:33:32.563 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:33:32.564 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:33:34.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:33 vm01.local ceph-mon[48597]: pgmap v2183: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:34.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:33 vm08.local ceph-mon[50506]: pgmap v2183: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:36.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:35 vm08.local ceph-mon[50506]: pgmap v2184: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:36.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:35 vm01.local ceph-mon[48597]: pgmap v2184: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:37.565 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:33:37.566 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:33:37.591 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:33:37.592 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:33:38.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:37 vm01.local ceph-mon[48597]: pgmap v2185: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:33:38.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:37 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:33:38.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:37 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:33:38.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:37 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:33:38.236 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:38 vm08.local ceph-mon[50506]: pgmap v2185: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:33:38.236 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:38 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:33:38.236 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:38 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:33:38.236 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:38 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:33:39.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:39 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:33:39.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:39 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:33:39.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:39 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:33:39.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:39 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:33:39.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:39 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:33:39.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:39 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:33:39.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:39 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:33:39.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:39 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:33:40.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:40 vm08.local ceph-mon[50506]: pgmap v2186: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:40.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:40 vm01.local ceph-mon[48597]: pgmap v2186: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:42 vm08.local ceph-mon[50506]: pgmap v2187: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:33:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:42 vm01.local ceph-mon[48597]: pgmap v2187: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:33:42.593 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:33:42.594 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:33:42.619 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:33:42.620 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:33:43.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:43 vm08.local ceph-mon[50506]: pgmap v2188: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:43.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:43 vm01.local ceph-mon[48597]: pgmap v2188: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:46.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:45 vm08.local ceph-mon[50506]: pgmap v2189: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:46.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:45 vm01.local ceph-mon[48597]: pgmap v2189: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:47.621 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:33:47.622 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:33:47.647 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:33:47.647 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:33:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:47 vm08.local ceph-mon[50506]: pgmap v2190: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:33:48.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:47 vm01.local ceph-mon[48597]: pgmap v2190: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:33:49.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:49 vm08.local ceph-mon[50506]: pgmap v2191: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:49.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:49 vm01.local ceph-mon[48597]: pgmap v2191: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:52.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:51 vm08.local ceph-mon[50506]: pgmap v2192: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:33:52.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:51 vm01.local ceph-mon[48597]: pgmap v2192: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:33:52.649 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:33:52.649 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:33:52.717 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:33:52.718 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:33:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:53 vm01.local ceph-mon[48597]: pgmap v2193: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:54.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:53 vm08.local ceph-mon[50506]: pgmap v2193: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:56.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:55 vm08.local ceph-mon[50506]: pgmap v2194: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:56.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:55 vm01.local ceph-mon[48597]: pgmap v2194: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:57.720 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:33:57.720 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:33:57.747 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:33:57.747 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:33:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:58 vm08.local ceph-mon[50506]: pgmap v2195: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:33:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:58 vm01.local ceph-mon[48597]: pgmap v2195: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:33:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:33:59 vm01.local ceph-mon[48597]: pgmap v2196: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:33:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:33:59 vm08.local ceph-mon[50506]: pgmap v2196: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:02.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:01 vm08.local ceph-mon[50506]: pgmap v2197: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:34:02.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:01 vm01.local ceph-mon[48597]: pgmap v2197: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:34:02.749 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:34:02.750 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:34:02.778 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:34:02.778 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:34:04.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:03 vm01.local ceph-mon[48597]: pgmap v2198: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:04.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:03 vm08.local ceph-mon[50506]: pgmap v2198: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:06.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:06 vm08.local ceph-mon[50506]: pgmap v2199: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:06.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:06 vm01.local ceph-mon[48597]: pgmap v2199: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:07.780 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:34:07.780 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:34:07.808 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:34:07.808 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:34:08.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:08 vm08.local ceph-mon[50506]: pgmap v2200: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:34:08.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:08 vm01.local ceph-mon[48597]: pgmap v2200: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:34:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:09 vm08.local ceph-mon[50506]: pgmap v2201: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:09 vm01.local ceph-mon[48597]: pgmap v2201: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:34:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:34:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:34:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:34:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:11 vm08.local ceph-mon[50506]: pgmap v2202: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:34:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:11 vm01.local ceph-mon[48597]: pgmap v2202: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:34:12.809 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:34:12.810 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:34:12.837 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:34:12.837 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:34:14.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:13 vm01.local ceph-mon[48597]: pgmap v2203: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:14.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:13 vm08.local ceph-mon[50506]: pgmap v2203: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:16.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:15 vm08.local ceph-mon[50506]: pgmap v2204: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:16.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:15 vm01.local ceph-mon[48597]: pgmap v2204: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:17 vm08.local ceph-mon[50506]: pgmap v2205: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:34:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:17 vm01.local ceph-mon[48597]: pgmap v2205: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:34:17.839 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:34:17.839 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:34:17.866 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:34:17.867 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:34:19.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:19 vm08.local ceph-mon[50506]: pgmap v2206: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:19.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:19 vm01.local ceph-mon[48597]: pgmap v2206: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:22.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:21 vm08.local ceph-mon[50506]: pgmap v2207: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:34:22.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:21 vm01.local ceph-mon[48597]: pgmap v2207: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:34:22.868 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:34:22.869 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:34:22.894 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:34:22.895 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:34:24.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:24 vm08.local ceph-mon[50506]: pgmap v2208: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:24.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:24 vm01.local ceph-mon[48597]: pgmap v2208: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:25.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:25 vm08.local ceph-mon[50506]: pgmap v2209: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:25.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:25 vm01.local ceph-mon[48597]: pgmap v2209: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:27.896 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:34:27.896 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:34:27.921 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:34:27.922 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:34:28.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:27 vm08.local ceph-mon[50506]: pgmap v2210: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:34:28.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:27 vm01.local ceph-mon[48597]: pgmap v2210: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:34:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:29 vm08.local ceph-mon[50506]: pgmap v2211: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:29 vm01.local ceph-mon[48597]: pgmap v2211: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:32.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:31 vm08.local ceph-mon[50506]: pgmap v2212: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:34:32.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:31 vm01.local ceph-mon[48597]: pgmap v2212: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:34:32.923 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:34:32.924 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:34:32.949 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:34:32.950 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:34:34.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:33 vm01.local ceph-mon[48597]: pgmap v2213: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:34.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:33 vm08.local ceph-mon[50506]: pgmap v2213: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:36.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:36 vm08.local ceph-mon[50506]: pgmap v2214: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:36.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:36 vm01.local ceph-mon[48597]: pgmap v2214: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:37.951 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:34:37.951 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:34:37.978 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:34:37.978 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:34:38.276 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:38 vm08.local ceph-mon[50506]: pgmap v2215: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:34:38.277 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:38 vm01.local ceph-mon[48597]: pgmap v2215: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:34:39.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:39 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:34:39.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:39 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:34:39.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:39 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:34:39.119 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:39 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:34:39.120 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:39 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:34:39.120 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:39 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:34:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:40 vm01.local ceph-mon[48597]: pgmap v2216: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:40 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:34:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:40 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:34:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:40 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:34:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:40 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:34:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:40 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:34:40.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:40 vm08.local ceph-mon[50506]: pgmap v2216: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:40.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:40 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:34:40.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:40 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:34:40.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:40 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:34:40.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:40 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:34:40.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:40 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:34:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:42 vm08.local ceph-mon[50506]: pgmap v2217: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:34:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:42 vm01.local ceph-mon[48597]: pgmap v2217: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:34:42.979 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:34:42.980 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:34:43.007 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:34:43.007 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:34:44.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:44 vm01.local ceph-mon[48597]: pgmap v2218: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:44.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:44 vm08.local ceph-mon[50506]: pgmap v2218: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:46.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:46 vm08.local ceph-mon[50506]: pgmap v2219: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:46.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:46 vm01.local ceph-mon[48597]: pgmap v2219: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:47.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:47 vm01.local ceph-mon[48597]: pgmap v2220: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:34:47.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:47 vm08.local ceph-mon[50506]: pgmap v2220: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:34:48.009 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:34:48.009 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:34:48.036 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:34:48.036 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:34:49.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:49 vm08.local ceph-mon[50506]: pgmap v2221: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:49.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:49 vm01.local ceph-mon[48597]: pgmap v2221: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:52.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:51 vm08.local ceph-mon[50506]: pgmap v2222: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:34:52.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:51 vm01.local ceph-mon[48597]: pgmap v2222: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:34:53.038 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:34:53.038 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:34:53.064 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:34:53.065 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:34:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:54 vm01.local ceph-mon[48597]: pgmap v2223: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:54.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:54 vm08.local ceph-mon[50506]: pgmap v2223: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:56.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:56 vm08.local ceph-mon[50506]: pgmap v2224: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:56.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:56 vm01.local ceph-mon[48597]: pgmap v2224: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:58.066 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:34:58.067 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:34:58.093 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:34:58.093 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:34:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:58 vm08.local ceph-mon[50506]: pgmap v2225: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:34:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:58 vm01.local ceph-mon[48597]: pgmap v2225: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:34:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:34:59 vm08.local ceph-mon[50506]: pgmap v2226: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:34:59.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:34:59 vm01.local ceph-mon[48597]: pgmap v2226: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:02.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:01 vm08.local ceph-mon[50506]: pgmap v2227: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:35:02.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:01 vm01.local ceph-mon[48597]: pgmap v2227: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:35:03.094 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:35:03.095 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:35:03.121 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:35:03.121 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:35:04.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:04 vm01.local ceph-mon[48597]: pgmap v2228: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:04.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:04 vm08.local ceph-mon[50506]: pgmap v2228: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:06.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:06 vm08.local ceph-mon[50506]: pgmap v2229: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:06.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:06 vm01.local ceph-mon[48597]: pgmap v2229: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:08.123 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:35:08.123 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:35:08.149 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:35:08.149 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:35:08.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:08 vm08.local ceph-mon[50506]: pgmap v2230: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:35:08.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:08 vm01.local ceph-mon[48597]: pgmap v2230: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:35:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:09 vm08.local ceph-mon[50506]: pgmap v2231: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:09 vm01.local ceph-mon[48597]: pgmap v2231: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:35:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:35:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:35:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:35:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:11 vm08.local ceph-mon[50506]: pgmap v2232: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:35:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:11 vm01.local ceph-mon[48597]: pgmap v2232: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:35:13.150 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:35:13.151 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:35:13.178 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:35:13.178 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:35:14.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:14 vm01.local ceph-mon[48597]: pgmap v2233: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:14.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:14 vm08.local ceph-mon[50506]: pgmap v2233: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:16.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:16 vm08.local ceph-mon[50506]: pgmap v2234: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:16.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:16 vm01.local ceph-mon[48597]: pgmap v2234: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:18.180 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:35:18.180 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:35:18.205 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:35:18.205 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:35:18.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:18 vm08.local ceph-mon[50506]: pgmap v2235: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:35:18.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:18 vm01.local ceph-mon[48597]: pgmap v2235: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:35:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:19 vm01.local ceph-mon[48597]: pgmap v2236: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:19.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:19 vm08.local ceph-mon[50506]: pgmap v2236: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:22.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:22 vm08.local ceph-mon[50506]: pgmap v2237: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:35:22.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:22 vm01.local ceph-mon[48597]: pgmap v2237: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:35:23.207 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:35:23.207 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:35:23.234 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:35:23.235 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:35:24.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:24 vm01.local ceph-mon[48597]: pgmap v2238: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:24.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:24 vm08.local ceph-mon[50506]: pgmap v2238: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:26.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:26 vm08.local ceph-mon[50506]: pgmap v2239: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:26.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:26 vm01.local ceph-mon[48597]: pgmap v2239: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:28.236 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:35:28.237 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:35:28.265 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:35:28.265 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:35:28.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:28 vm08.local ceph-mon[50506]: pgmap v2240: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:35:28.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:28 vm01.local ceph-mon[48597]: pgmap v2240: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:35:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:29 vm08.local ceph-mon[50506]: pgmap v2241: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:29 vm01.local ceph-mon[48597]: pgmap v2241: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:32.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:32 vm08.local ceph-mon[50506]: pgmap v2242: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:35:32.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:32 vm01.local ceph-mon[48597]: pgmap v2242: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:35:33.266 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:35:33.267 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:35:33.293 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:35:33.293 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:35:34.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:34 vm01.local ceph-mon[48597]: pgmap v2243: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:34.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:34 vm08.local ceph-mon[50506]: pgmap v2243: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:36.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:36 vm08.local ceph-mon[50506]: pgmap v2244: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:36.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:36 vm01.local ceph-mon[48597]: pgmap v2244: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:38.295 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:35:38.295 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:35:38.321 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:35:38.321 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:35:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:38 vm08.local ceph-mon[50506]: pgmap v2245: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:35:38.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:38 vm01.local ceph-mon[48597]: pgmap v2245: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:35:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:39 vm08.local ceph-mon[50506]: pgmap v2246: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:39.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:39 vm01.local ceph-mon[48597]: pgmap v2246: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:40 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:35:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:40 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:35:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:40 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:35:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:40 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:35:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:40 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:35:40.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:40 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:35:40.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:40 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:35:40.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:40 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:35:40.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:40 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:35:40.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:40 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:35:41.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:41 vm08.local ceph-mon[50506]: pgmap v2247: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:35:41.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:41 vm01.local ceph-mon[48597]: pgmap v2247: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:35:43.323 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:35:43.323 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:35:43.349 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:35:43.349 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:35:44.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:44 vm01.local ceph-mon[48597]: pgmap v2248: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:44.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:44 vm08.local ceph-mon[50506]: pgmap v2248: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:46.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:46 vm08.local ceph-mon[50506]: pgmap v2249: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:46.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:46 vm01.local ceph-mon[48597]: pgmap v2249: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:48 vm08.local ceph-mon[50506]: pgmap v2250: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:35:48.350 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:35:48.351 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:35:48.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:48 vm01.local ceph-mon[48597]: pgmap v2250: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:35:48.377 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:35:48.378 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:35:49.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:49 vm01.local ceph-mon[48597]: pgmap v2251: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:49.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:49 vm08.local ceph-mon[50506]: pgmap v2251: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:52.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:52 vm08.local ceph-mon[50506]: pgmap v2252: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:35:52.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:52 vm01.local ceph-mon[48597]: pgmap v2252: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:35:53.379 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:35:53.380 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:35:53.405 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:35:53.405 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:35:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:54 vm01.local ceph-mon[48597]: pgmap v2253: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:54.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:54 vm08.local ceph-mon[50506]: pgmap v2253: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:56.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:56 vm08.local ceph-mon[50506]: pgmap v2254: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:56.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:56 vm01.local ceph-mon[48597]: pgmap v2254: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:58 vm08.local ceph-mon[50506]: pgmap v2255: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:35:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:58 vm01.local ceph-mon[48597]: pgmap v2255: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:35:58.407 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:35:58.407 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:35:58.433 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:35:58.433 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:35:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:35:59 vm08.local ceph-mon[50506]: pgmap v2256: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:35:59.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:35:59 vm01.local ceph-mon[48597]: pgmap v2256: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:02.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:02 vm08.local ceph-mon[50506]: pgmap v2257: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:36:02.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:02 vm01.local ceph-mon[48597]: pgmap v2257: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:36:03.434 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:36:03.435 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:36:03.462 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:36:03.463 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:36:04.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:04 vm01.local ceph-mon[48597]: pgmap v2258: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:04.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:04 vm08.local ceph-mon[50506]: pgmap v2258: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:06.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:06 vm08.local ceph-mon[50506]: pgmap v2259: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:06.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:06 vm01.local ceph-mon[48597]: pgmap v2259: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:08.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:08 vm08.local ceph-mon[50506]: pgmap v2260: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:36:08.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:08 vm01.local ceph-mon[48597]: pgmap v2260: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:36:08.464 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:36:08.464 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:36:08.490 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:36:08.491 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:36:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:09 vm08.local ceph-mon[50506]: pgmap v2261: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:09 vm01.local ceph-mon[48597]: pgmap v2261: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:36:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:36:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:36:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:36:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:11 vm08.local ceph-mon[50506]: pgmap v2262: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:36:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:11 vm01.local ceph-mon[48597]: pgmap v2262: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:36:13.492 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:36:13.492 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:36:13.518 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:36:13.519 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:36:14.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:14 vm01.local ceph-mon[48597]: pgmap v2263: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:14.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:14 vm08.local ceph-mon[50506]: pgmap v2263: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:16.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:16 vm08.local ceph-mon[50506]: pgmap v2264: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:16.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:16 vm01.local ceph-mon[48597]: pgmap v2264: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:18.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:18 vm08.local ceph-mon[50506]: pgmap v2265: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:36:18.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:18 vm01.local ceph-mon[48597]: pgmap v2265: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:36:18.520 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:36:18.520 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:36:18.546 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:36:18.547 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:36:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:19 vm01.local ceph-mon[48597]: pgmap v2266: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:19.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:19 vm08.local ceph-mon[50506]: pgmap v2266: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:22.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:22 vm08.local ceph-mon[50506]: pgmap v2267: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:36:22.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:22 vm01.local ceph-mon[48597]: pgmap v2267: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:36:23.548 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:36:23.548 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:36:23.574 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:36:23.574 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:36:24.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:24 vm01.local ceph-mon[48597]: pgmap v2268: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:24.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:24 vm08.local ceph-mon[50506]: pgmap v2268: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:26.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:26 vm08.local ceph-mon[50506]: pgmap v2269: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:26.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:26 vm01.local ceph-mon[48597]: pgmap v2269: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:28.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:28 vm08.local ceph-mon[50506]: pgmap v2270: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:36:28.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:28 vm01.local ceph-mon[48597]: pgmap v2270: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:36:28.576 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:36:28.576 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:36:28.601 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:36:28.601 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:36:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:29 vm08.local ceph-mon[50506]: pgmap v2271: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:29 vm01.local ceph-mon[48597]: pgmap v2271: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:32.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:32 vm08.local ceph-mon[50506]: pgmap v2272: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:36:32.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:32 vm01.local ceph-mon[48597]: pgmap v2272: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:36:33.603 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:36:33.603 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:36:33.629 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:36:33.629 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:36:34.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:34 vm01.local ceph-mon[48597]: pgmap v2273: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:34.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:34 vm08.local ceph-mon[50506]: pgmap v2273: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:36.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:36 vm08.local ceph-mon[50506]: pgmap v2274: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:36.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:36 vm01.local ceph-mon[48597]: pgmap v2274: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:38 vm08.local ceph-mon[50506]: pgmap v2275: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:36:38.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:38 vm01.local ceph-mon[48597]: pgmap v2275: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:36:38.631 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:36:38.631 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:36:38.656 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:36:38.657 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:36:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:39 vm08.local ceph-mon[50506]: pgmap v2276: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:39.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:39 vm01.local ceph-mon[48597]: pgmap v2276: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:40 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:36:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:40 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:36:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:40 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:36:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:40 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:36:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:40 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:36:40.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:40 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:36:40.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:40 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:36:40.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:40 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:36:40.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:40 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:36:40.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:40 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:36:41.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:41 vm08.local ceph-mon[50506]: pgmap v2277: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:36:41.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:41 vm01.local ceph-mon[48597]: pgmap v2277: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:36:43.658 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:36:43.658 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:36:43.684 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:36:43.684 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:36:44.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:44 vm01.local ceph-mon[48597]: pgmap v2278: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:44.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:44 vm08.local ceph-mon[50506]: pgmap v2278: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:46.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:46 vm08.local ceph-mon[50506]: pgmap v2279: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:46.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:46 vm01.local ceph-mon[48597]: pgmap v2279: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:48 vm08.local ceph-mon[50506]: pgmap v2280: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:36:48.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:48 vm01.local ceph-mon[48597]: pgmap v2280: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:36:48.686 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:36:48.686 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:36:48.712 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:36:48.712 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:36:49.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:49 vm01.local ceph-mon[48597]: pgmap v2281: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:49.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:49 vm08.local ceph-mon[50506]: pgmap v2281: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:52.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:52 vm08.local ceph-mon[50506]: pgmap v2282: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:36:52.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:52 vm01.local ceph-mon[48597]: pgmap v2282: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:36:53.714 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:36:53.714 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:36:53.741 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:36:53.742 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:36:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:54 vm01.local ceph-mon[48597]: pgmap v2283: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:54.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:54 vm08.local ceph-mon[50506]: pgmap v2283: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:56.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:56 vm08.local ceph-mon[50506]: pgmap v2284: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:56.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:56 vm01.local ceph-mon[48597]: pgmap v2284: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:58 vm08.local ceph-mon[50506]: pgmap v2285: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:36:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:58 vm01.local ceph-mon[48597]: pgmap v2285: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:36:58.743 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:36:58.743 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:36:58.769 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:36:58.770 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:36:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:36:59 vm08.local ceph-mon[50506]: pgmap v2286: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:36:59.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:36:59 vm01.local ceph-mon[48597]: pgmap v2286: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:02.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:02 vm08.local ceph-mon[50506]: pgmap v2287: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:37:02.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:02 vm01.local ceph-mon[48597]: pgmap v2287: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:37:03.771 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:37:03.772 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:37:03.797 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:37:03.798 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:37:04.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:04 vm01.local ceph-mon[48597]: pgmap v2288: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:04.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:04 vm08.local ceph-mon[50506]: pgmap v2288: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:06.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:06 vm08.local ceph-mon[50506]: pgmap v2289: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:06.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:06 vm01.local ceph-mon[48597]: pgmap v2289: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:08.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:08 vm08.local ceph-mon[50506]: pgmap v2290: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:37:08.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:08 vm01.local ceph-mon[48597]: pgmap v2290: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:37:08.799 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:37:08.799 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:37:08.825 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:37:08.825 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:37:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:09 vm08.local ceph-mon[50506]: pgmap v2291: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:09 vm01.local ceph-mon[48597]: pgmap v2291: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:37:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:37:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:37:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:37:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:11 vm08.local ceph-mon[50506]: pgmap v2292: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:37:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:11 vm01.local ceph-mon[48597]: pgmap v2292: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:37:13.827 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:37:13.827 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:37:13.853 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:37:13.853 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:37:14.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:14 vm01.local ceph-mon[48597]: pgmap v2293: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:14.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:14 vm08.local ceph-mon[50506]: pgmap v2293: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:16.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:16 vm08.local ceph-mon[50506]: pgmap v2294: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:16.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:16 vm01.local ceph-mon[48597]: pgmap v2294: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:18.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:18 vm08.local ceph-mon[50506]: pgmap v2295: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:37:18.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:18 vm01.local ceph-mon[48597]: pgmap v2295: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:37:18.855 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:37:18.855 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:37:18.881 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:37:18.881 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:37:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:19 vm01.local ceph-mon[48597]: pgmap v2296: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:19.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:19 vm08.local ceph-mon[50506]: pgmap v2296: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:22.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:22 vm08.local ceph-mon[50506]: pgmap v2297: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:37:22.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:22 vm01.local ceph-mon[48597]: pgmap v2297: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:37:23.883 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:37:23.883 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:37:23.909 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:37:23.909 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:37:24.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:24 vm01.local ceph-mon[48597]: pgmap v2298: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:24.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:24 vm08.local ceph-mon[50506]: pgmap v2298: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:26.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:26 vm08.local ceph-mon[50506]: pgmap v2299: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:26.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:26 vm01.local ceph-mon[48597]: pgmap v2299: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:28.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:28 vm08.local ceph-mon[50506]: pgmap v2300: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:37:28.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:28 vm01.local ceph-mon[48597]: pgmap v2300: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:37:28.911 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:37:28.911 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:37:28.936 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:37:28.937 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:37:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:29 vm08.local ceph-mon[50506]: pgmap v2301: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:29 vm01.local ceph-mon[48597]: pgmap v2301: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:32.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:32 vm08.local ceph-mon[50506]: pgmap v2302: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:37:32.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:32 vm01.local ceph-mon[48597]: pgmap v2302: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:37:33.938 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:37:33.938 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:37:33.964 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:37:33.964 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:37:34.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:34 vm08.local ceph-mon[50506]: pgmap v2303: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:34.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:34 vm01.local ceph-mon[48597]: pgmap v2303: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:36.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:36 vm08.local ceph-mon[50506]: pgmap v2304: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:36.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:36 vm01.local ceph-mon[48597]: pgmap v2304: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:38 vm08.local ceph-mon[50506]: pgmap v2305: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:37:38.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:38 vm01.local ceph-mon[48597]: pgmap v2305: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:37:38.965 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:37:38.966 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:37:38.993 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:37:38.994 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:37:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:39 vm08.local ceph-mon[50506]: pgmap v2306: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:39.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:39 vm01.local ceph-mon[48597]: pgmap v2306: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:40.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:40 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:37:40.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:40 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:37:40.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:40 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:37:40.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:40 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:37:40.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:40 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:37:40.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:40 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:37:41.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:41 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:37:41.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:41 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:37:41.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:41 vm08.local ceph-mon[50506]: pgmap v2307: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:37:41.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:41 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:37:41.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:41 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:37:41.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:41 vm01.local ceph-mon[48597]: pgmap v2307: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:37:43.995 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:37:43.996 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:37:44.023 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:37:44.023 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:37:44.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:44 vm08.local ceph-mon[50506]: pgmap v2308: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:44.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:44 vm01.local ceph-mon[48597]: pgmap v2308: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:46.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:46 vm08.local ceph-mon[50506]: pgmap v2309: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:46.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:46 vm01.local ceph-mon[48597]: pgmap v2309: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:48 vm08.local ceph-mon[50506]: pgmap v2310: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:37:48.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:48 vm01.local ceph-mon[48597]: pgmap v2310: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:37:49.024 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:37:49.025 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:37:49.050 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:37:49.051 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:37:49.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:49 vm01.local ceph-mon[48597]: pgmap v2311: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:49.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:49 vm08.local ceph-mon[50506]: pgmap v2311: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:52.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:52 vm08.local ceph-mon[50506]: pgmap v2312: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:37:52.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:52 vm01.local ceph-mon[48597]: pgmap v2312: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:37:54.052 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:37:54.052 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:37:54.088 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:37:54.089 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:37:54.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:54 vm08.local ceph-mon[50506]: pgmap v2313: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:54.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:54 vm01.local ceph-mon[48597]: pgmap v2313: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:56.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:56 vm08.local ceph-mon[50506]: pgmap v2314: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:56.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:56 vm01.local ceph-mon[48597]: pgmap v2314: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:58 vm08.local ceph-mon[50506]: pgmap v2315: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:37:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:58 vm01.local ceph-mon[48597]: pgmap v2315: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:37:59.090 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:37:59.091 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:37:59.116 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:37:59.117 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:37:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:37:59 vm08.local ceph-mon[50506]: pgmap v2316: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:37:59.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:37:59 vm01.local ceph-mon[48597]: pgmap v2316: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:02.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:02 vm08.local ceph-mon[50506]: pgmap v2317: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:38:02.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:02 vm01.local ceph-mon[48597]: pgmap v2317: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:38:04.118 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:38:04.118 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:38:04.144 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:38:04.144 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:38:04.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:04 vm08.local ceph-mon[50506]: pgmap v2318: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:04.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:04 vm01.local ceph-mon[48597]: pgmap v2318: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:06.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:06 vm08.local ceph-mon[50506]: pgmap v2319: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:06.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:06 vm01.local ceph-mon[48597]: pgmap v2319: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:08.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:08 vm08.local ceph-mon[50506]: pgmap v2320: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:38:08.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:08 vm01.local ceph-mon[48597]: pgmap v2320: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:38:09.146 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:38:09.146 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:38:09.172 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:38:09.172 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:38:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:09 vm08.local ceph-mon[50506]: pgmap v2321: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:09 vm01.local ceph-mon[48597]: pgmap v2321: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:38:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:38:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:38:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:38:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:11 vm08.local ceph-mon[50506]: pgmap v2322: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:38:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:11 vm01.local ceph-mon[48597]: pgmap v2322: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:38:14.173 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:38:14.174 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:38:14.200 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:38:14.200 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:38:14.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:14 vm08.local ceph-mon[50506]: pgmap v2323: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:14.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:14 vm01.local ceph-mon[48597]: pgmap v2323: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:16.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:16 vm08.local ceph-mon[50506]: pgmap v2324: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:16.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:16 vm01.local ceph-mon[48597]: pgmap v2324: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:18.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:18 vm08.local ceph-mon[50506]: pgmap v2325: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:38:18.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:18 vm01.local ceph-mon[48597]: pgmap v2325: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:38:19.201 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:38:19.202 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:38:19.229 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:38:19.230 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:38:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:19 vm01.local ceph-mon[48597]: pgmap v2326: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:19.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:19 vm08.local ceph-mon[50506]: pgmap v2326: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:22.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:22 vm08.local ceph-mon[50506]: pgmap v2327: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:38:22.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:22 vm01.local ceph-mon[48597]: pgmap v2327: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:38:24.231 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:38:24.231 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:38:24.258 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:38:24.258 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:38:24.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:24 vm08.local ceph-mon[50506]: pgmap v2328: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:24.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:24 vm01.local ceph-mon[48597]: pgmap v2328: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:26.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:26 vm08.local ceph-mon[50506]: pgmap v2329: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:26.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:26 vm01.local ceph-mon[48597]: pgmap v2329: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:28.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:28 vm08.local ceph-mon[50506]: pgmap v2330: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:38:28.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:28 vm01.local ceph-mon[48597]: pgmap v2330: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:38:29.259 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:38:29.260 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:38:29.286 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:38:29.286 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:38:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:29 vm08.local ceph-mon[50506]: pgmap v2331: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:29 vm01.local ceph-mon[48597]: pgmap v2331: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:32.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:32 vm08.local ceph-mon[50506]: pgmap v2332: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:38:32.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:32 vm01.local ceph-mon[48597]: pgmap v2332: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:38:34.288 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:38:34.288 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:38:34.314 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:38:34.314 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:38:34.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:34 vm08.local ceph-mon[50506]: pgmap v2333: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:34.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:34 vm01.local ceph-mon[48597]: pgmap v2333: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:36.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:36 vm08.local ceph-mon[50506]: pgmap v2334: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:36.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:36 vm01.local ceph-mon[48597]: pgmap v2334: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:38 vm08.local ceph-mon[50506]: pgmap v2335: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:38:38.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:38 vm01.local ceph-mon[48597]: pgmap v2335: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:38:39.316 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:38:39.316 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:38:39.342 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:38:39.342 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:38:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:39 vm08.local ceph-mon[50506]: pgmap v2336: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:39.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:39 vm01.local ceph-mon[48597]: pgmap v2336: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:40 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:38:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:40 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:38:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:40 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:38:40.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:40 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:38:40.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:40 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:38:40.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:40 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:38:42.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:41 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:38:42.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:41 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:38:42.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:41 vm08.local ceph-mon[50506]: pgmap v2337: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:38:42.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:41 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:38:42.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:41 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:38:42.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:41 vm01.local ceph-mon[48597]: pgmap v2337: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:38:44.343 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:38:44.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:44 vm08.local ceph-mon[50506]: pgmap v2338: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:44.344 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:38:44.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:44 vm01.local ceph-mon[48597]: pgmap v2338: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:44.370 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:38:44.370 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:38:46.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:46 vm08.local ceph-mon[50506]: pgmap v2339: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:46.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:46 vm01.local ceph-mon[48597]: pgmap v2339: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:48 vm08.local ceph-mon[50506]: pgmap v2340: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:38:48.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:48 vm01.local ceph-mon[48597]: pgmap v2340: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:38:49.372 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:38:49.372 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:38:49.397 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:38:49.397 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:38:49.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:49 vm01.local ceph-mon[48597]: pgmap v2341: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:49.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:49 vm08.local ceph-mon[50506]: pgmap v2341: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:52.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:52 vm08.local ceph-mon[50506]: pgmap v2342: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:38:52.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:52 vm01.local ceph-mon[48597]: pgmap v2342: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:38:54.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:54 vm08.local ceph-mon[50506]: pgmap v2343: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:54.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:54 vm01.local ceph-mon[48597]: pgmap v2343: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:54.399 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:38:54.399 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:38:54.427 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:38:54.427 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:38:56.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:56 vm08.local ceph-mon[50506]: pgmap v2344: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:56.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:56 vm01.local ceph-mon[48597]: pgmap v2344: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:58 vm08.local ceph-mon[50506]: pgmap v2345: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:38:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:58 vm01.local ceph-mon[48597]: pgmap v2345: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:38:59.428 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:38:59.429 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:38:59.455 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:38:59.455 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:38:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:38:59 vm08.local ceph-mon[50506]: pgmap v2346: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:38:59.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:38:59 vm01.local ceph-mon[48597]: pgmap v2346: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:02.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:02 vm08.local ceph-mon[50506]: pgmap v2347: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:39:02.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:02 vm01.local ceph-mon[48597]: pgmap v2347: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:39:04.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:04 vm08.local ceph-mon[50506]: pgmap v2348: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:04.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:04 vm01.local ceph-mon[48597]: pgmap v2348: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:04.456 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:39:04.457 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:39:04.486 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:39:04.486 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:39:06.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:06 vm08.local ceph-mon[50506]: pgmap v2349: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:06.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:06 vm01.local ceph-mon[48597]: pgmap v2349: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:08.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:08 vm08.local ceph-mon[50506]: pgmap v2350: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:39:08.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:08 vm01.local ceph-mon[48597]: pgmap v2350: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:39:09.487 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:39:09.488 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:39:09.516 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:39:09.516 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:39:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:09 vm08.local ceph-mon[50506]: pgmap v2351: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:09 vm01.local ceph-mon[48597]: pgmap v2351: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:39:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:39:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:39:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:39:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:11 vm08.local ceph-mon[50506]: pgmap v2352: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:39:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:11 vm01.local ceph-mon[48597]: pgmap v2352: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:39:14.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:14 vm08.local ceph-mon[50506]: pgmap v2353: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:14.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:14 vm01.local ceph-mon[48597]: pgmap v2353: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:14.517 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:39:14.518 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:39:14.543 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:39:14.543 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:39:16.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:16 vm08.local ceph-mon[50506]: pgmap v2354: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:16.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:16 vm01.local ceph-mon[48597]: pgmap v2354: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:18.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:18 vm08.local ceph-mon[50506]: pgmap v2355: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:39:18.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:18 vm01.local ceph-mon[48597]: pgmap v2355: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:39:19.545 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:39:19.545 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:39:19.570 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:39:19.571 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:39:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:19 vm01.local ceph-mon[48597]: pgmap v2356: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:19.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:19 vm08.local ceph-mon[50506]: pgmap v2356: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:22.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:22 vm08.local ceph-mon[50506]: pgmap v2357: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:39:22.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:22 vm01.local ceph-mon[48597]: pgmap v2357: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:39:24.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:24 vm08.local ceph-mon[50506]: pgmap v2358: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:24.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:24 vm01.local ceph-mon[48597]: pgmap v2358: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:24.572 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:39:24.573 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:39:24.599 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:39:24.599 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:39:26.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:26 vm08.local ceph-mon[50506]: pgmap v2359: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:26.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:26 vm01.local ceph-mon[48597]: pgmap v2359: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:28.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:28 vm08.local ceph-mon[50506]: pgmap v2360: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:39:28.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:28 vm01.local ceph-mon[48597]: pgmap v2360: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:39:29.601 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:39:29.601 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:39:29.626 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:39:29.627 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:39:29.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:29 vm08.local ceph-mon[50506]: pgmap v2361: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:29 vm01.local ceph-mon[48597]: pgmap v2361: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:32.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:32 vm08.local ceph-mon[50506]: pgmap v2362: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:39:32.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:32 vm01.local ceph-mon[48597]: pgmap v2362: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:39:34.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:34 vm08.local ceph-mon[50506]: pgmap v2363: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:34.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:34 vm01.local ceph-mon[48597]: pgmap v2363: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:34.628 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:39:34.628 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:39:34.653 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:39:34.654 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:39:36.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:36 vm08.local ceph-mon[50506]: pgmap v2364: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:36.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:36 vm01.local ceph-mon[48597]: pgmap v2364: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:38 vm08.local ceph-mon[50506]: pgmap v2365: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:39:38.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:38 vm01.local ceph-mon[48597]: pgmap v2365: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:39:39.655 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:39:39.656 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:39:39.684 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:39:39.684 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:39:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:39 vm08.local ceph-mon[50506]: pgmap v2366: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:39.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:39 vm01.local ceph-mon[48597]: pgmap v2366: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:41.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:40 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:39:41.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:40 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:39:41.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:40 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:39:41.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:40 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:39:41.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:40 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:39:41.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:40 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:39:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:41 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:39:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:41 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:39:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:41 vm08.local ceph-mon[50506]: pgmap v2367: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:39:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:41 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:39:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:41 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:39:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:41 vm01.local ceph-mon[48597]: pgmap v2367: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:39:44.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:44 vm08.local ceph-mon[50506]: pgmap v2368: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:44.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:44 vm01.local ceph-mon[48597]: pgmap v2368: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:44.685 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:39:44.686 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:39:44.713 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:39:44.714 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:39:46.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:46 vm08.local ceph-mon[50506]: pgmap v2369: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:46.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:46 vm01.local ceph-mon[48597]: pgmap v2369: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:48 vm08.local ceph-mon[50506]: pgmap v2370: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:39:48.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:48 vm01.local ceph-mon[48597]: pgmap v2370: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:39:49.715 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:39:49.715 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:39:49.742 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:39:49.743 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:39:49.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:49 vm08.local ceph-mon[50506]: pgmap v2371: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:49.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:49 vm01.local ceph-mon[48597]: pgmap v2371: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:52.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:52 vm08.local ceph-mon[50506]: pgmap v2372: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:39:52.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:52 vm01.local ceph-mon[48597]: pgmap v2372: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:39:54.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:54 vm08.local ceph-mon[50506]: pgmap v2373: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:54.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:54 vm01.local ceph-mon[48597]: pgmap v2373: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:54.744 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:39:54.744 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:39:54.771 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:39:54.772 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:39:56.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:56 vm08.local ceph-mon[50506]: pgmap v2374: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:56.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:56 vm01.local ceph-mon[48597]: pgmap v2374: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:58 vm08.local ceph-mon[50506]: pgmap v2375: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:39:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:58 vm01.local ceph-mon[48597]: pgmap v2375: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:39:59.773 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:39:59.773 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:39:59.801 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:39:59.801 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:39:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:39:59 vm08.local ceph-mon[50506]: pgmap v2376: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:39:59.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:39:59 vm01.local ceph-mon[48597]: pgmap v2376: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:00.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:00 vm08.local ceph-mon[50506]: overall HEALTH_OK 2026-03-10T11:40:00.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:00 vm01.local ceph-mon[48597]: overall HEALTH_OK 2026-03-10T11:40:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:01 vm08.local ceph-mon[50506]: pgmap v2377: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:40:01.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:01 vm01.local ceph-mon[48597]: pgmap v2377: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:40:04.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:04 vm08.local ceph-mon[50506]: pgmap v2378: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:04.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:04 vm01.local ceph-mon[48597]: pgmap v2378: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:04.803 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:40:04.803 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:40:04.832 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:40:04.832 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:40:06.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:06 vm08.local ceph-mon[50506]: pgmap v2379: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:06.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:06 vm01.local ceph-mon[48597]: pgmap v2379: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:08.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:08 vm08.local ceph-mon[50506]: pgmap v2380: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:40:08.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:08 vm01.local ceph-mon[48597]: pgmap v2380: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:40:09.833 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:40:09.834 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:40:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:09 vm08.local ceph-mon[50506]: pgmap v2381: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:09.860 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:40:09.860 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:40:09.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:09 vm01.local ceph-mon[48597]: pgmap v2381: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:40:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:40:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:40:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:40:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:11 vm08.local ceph-mon[50506]: pgmap v2382: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:40:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:11 vm01.local ceph-mon[48597]: pgmap v2382: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:40:14.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:14 vm08.local ceph-mon[50506]: pgmap v2383: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:14.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:14 vm01.local ceph-mon[48597]: pgmap v2383: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:14.862 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:40:14.862 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:40:14.888 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:40:14.888 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:40:16.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:16 vm08.local ceph-mon[50506]: pgmap v2384: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:16.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:16 vm01.local ceph-mon[48597]: pgmap v2384: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:18.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:18 vm08.local ceph-mon[50506]: pgmap v2385: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:40:18.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:18 vm01.local ceph-mon[48597]: pgmap v2385: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:40:19.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:19 vm08.local ceph-mon[50506]: pgmap v2386: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:19.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:19 vm01.local ceph-mon[48597]: pgmap v2386: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:19.890 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:40:19.890 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:40:19.916 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:40:19.917 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:40:22.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:22 vm08.local ceph-mon[50506]: pgmap v2387: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:40:22.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:22 vm01.local ceph-mon[48597]: pgmap v2387: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:40:24.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:24 vm08.local ceph-mon[50506]: pgmap v2388: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:24.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:24 vm01.local ceph-mon[48597]: pgmap v2388: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:24.918 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:40:24.918 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:40:24.946 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:40:24.946 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:40:26.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:26 vm08.local ceph-mon[50506]: pgmap v2389: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:26.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:26 vm01.local ceph-mon[48597]: pgmap v2389: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:28.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:28 vm08.local ceph-mon[50506]: pgmap v2390: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:40:28.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:28 vm01.local ceph-mon[48597]: pgmap v2390: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:40:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:29 vm08.local ceph-mon[50506]: pgmap v2391: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:29 vm01.local ceph-mon[48597]: pgmap v2391: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:29.948 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:40:29.948 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:40:29.974 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:40:29.975 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:40:32.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:32 vm08.local ceph-mon[50506]: pgmap v2392: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:40:32.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:32 vm01.local ceph-mon[48597]: pgmap v2392: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:40:34.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:34 vm08.local ceph-mon[50506]: pgmap v2393: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:34.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:34 vm01.local ceph-mon[48597]: pgmap v2393: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:34.976 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:40:34.977 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:40:35.003 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:40:35.004 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:40:36.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:36 vm08.local ceph-mon[50506]: pgmap v2394: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:36.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:36 vm01.local ceph-mon[48597]: pgmap v2394: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:38.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:38 vm08.local ceph-mon[50506]: pgmap v2395: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:40:38.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:38 vm01.local ceph-mon[48597]: pgmap v2395: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:40:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:39 vm08.local ceph-mon[50506]: pgmap v2396: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:39.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:39 vm01.local ceph-mon[48597]: pgmap v2396: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:40.005 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:40:40.005 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:40:40.032 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:40:40.032 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:40:41.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:41 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:40:41.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:41 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:40:41.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:41 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:40:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:41 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:40:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:41 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:40:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:41 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:40:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:42 vm08.local ceph-mon[50506]: pgmap v2397: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:40:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:42 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:40:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:42 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:40:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:42 vm01.local ceph-mon[48597]: pgmap v2397: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:40:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:42 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:40:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:42 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:40:44.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:44 vm08.local ceph-mon[50506]: pgmap v2398: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:44.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:44 vm01.local ceph-mon[48597]: pgmap v2398: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:45.033 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:40:45.034 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:40:45.060 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:40:45.060 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:40:46.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:46 vm08.local ceph-mon[50506]: pgmap v2399: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:46.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:46 vm01.local ceph-mon[48597]: pgmap v2399: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:48.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:48 vm08.local ceph-mon[50506]: pgmap v2400: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:40:48.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:48 vm01.local ceph-mon[48597]: pgmap v2400: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:40:49.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:49 vm08.local ceph-mon[50506]: pgmap v2401: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:49.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:49 vm01.local ceph-mon[48597]: pgmap v2401: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:50.062 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:40:50.062 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:40:50.087 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:40:50.088 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:40:52.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:52 vm08.local ceph-mon[50506]: pgmap v2402: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:40:52.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:52 vm01.local ceph-mon[48597]: pgmap v2402: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:40:54.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:54 vm08.local ceph-mon[50506]: pgmap v2403: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:54.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:54 vm01.local ceph-mon[48597]: pgmap v2403: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:55.089 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:40:55.089 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:40:55.115 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:40:55.115 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:40:56.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:56 vm08.local ceph-mon[50506]: pgmap v2404: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:56.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:56 vm01.local ceph-mon[48597]: pgmap v2404: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:58 vm08.local ceph-mon[50506]: pgmap v2405: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:40:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:58 vm01.local ceph-mon[48597]: pgmap v2405: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:40:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:40:59 vm08.local ceph-mon[50506]: pgmap v2406: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:40:59.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:40:59 vm01.local ceph-mon[48597]: pgmap v2406: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:00.117 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:41:00.117 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:41:00.144 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:41:00.144 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:41:02.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:02 vm08.local ceph-mon[50506]: pgmap v2407: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:41:02.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:02 vm01.local ceph-mon[48597]: pgmap v2407: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:41:04.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:04 vm08.local ceph-mon[50506]: pgmap v2408: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:04.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:04 vm01.local ceph-mon[48597]: pgmap v2408: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:05.145 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:41:05.145 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:41:05.173 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:41:05.174 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:41:06.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:06 vm08.local ceph-mon[50506]: pgmap v2409: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:06.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:06 vm01.local ceph-mon[48597]: pgmap v2409: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:08.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:08 vm08.local ceph-mon[50506]: pgmap v2410: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:41:08.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:08 vm01.local ceph-mon[48597]: pgmap v2410: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:41:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:09 vm08.local ceph-mon[50506]: pgmap v2411: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:09 vm01.local ceph-mon[48597]: pgmap v2411: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:10.175 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:41:10.175 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:41:10.202 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:41:10.203 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:41:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:41:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:41:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:41:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:41:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:11 vm08.local ceph-mon[50506]: pgmap v2412: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:41:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:11 vm01.local ceph-mon[48597]: pgmap v2412: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:41:14.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:14 vm08.local ceph-mon[50506]: pgmap v2413: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:14.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:14 vm01.local ceph-mon[48597]: pgmap v2413: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:15.204 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:41:15.205 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:41:15.232 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:41:15.233 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:41:16.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:16 vm08.local ceph-mon[50506]: pgmap v2414: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:16.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:16 vm01.local ceph-mon[48597]: pgmap v2414: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:18.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:18 vm08.local ceph-mon[50506]: pgmap v2415: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:41:18.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:18 vm01.local ceph-mon[48597]: pgmap v2415: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:41:19.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:19 vm08.local ceph-mon[50506]: pgmap v2416: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:19.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:19 vm01.local ceph-mon[48597]: pgmap v2416: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:20.234 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:41:20.235 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:41:20.261 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:41:20.262 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:41:22.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:22 vm08.local ceph-mon[50506]: pgmap v2417: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:41:22.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:22 vm01.local ceph-mon[48597]: pgmap v2417: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:41:24.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:24 vm08.local ceph-mon[50506]: pgmap v2418: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:24.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:24 vm01.local ceph-mon[48597]: pgmap v2418: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:25.263 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:41:25.264 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:41:25.289 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:41:25.289 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:41:26.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:26 vm08.local ceph-mon[50506]: pgmap v2419: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:26.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:26 vm01.local ceph-mon[48597]: pgmap v2419: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:28.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:28 vm08.local ceph-mon[50506]: pgmap v2420: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:41:28.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:28 vm01.local ceph-mon[48597]: pgmap v2420: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:41:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:29 vm08.local ceph-mon[50506]: pgmap v2421: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:29 vm01.local ceph-mon[48597]: pgmap v2421: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:30.290 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:41:30.291 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:41:30.317 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:41:30.317 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:41:32.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:32 vm08.local ceph-mon[50506]: pgmap v2422: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:41:32.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:32 vm01.local ceph-mon[48597]: pgmap v2422: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:41:34.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:34 vm08.local ceph-mon[50506]: pgmap v2423: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:34.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:34 vm01.local ceph-mon[48597]: pgmap v2423: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:35.319 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:41:35.319 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:41:35.346 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:41:35.346 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:41:36.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:36 vm08.local ceph-mon[50506]: pgmap v2424: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:36.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:36 vm01.local ceph-mon[48597]: pgmap v2424: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:38.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:38 vm01.local ceph-mon[48597]: pgmap v2425: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:41:38.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:38 vm08.local ceph-mon[50506]: pgmap v2425: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:41:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:39 vm08.local ceph-mon[50506]: pgmap v2426: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:39.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:39 vm01.local ceph-mon[48597]: pgmap v2426: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:40.347 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:41:40.348 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:41:40.374 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:41:40.375 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:41:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:42 vm08.local ceph-mon[50506]: pgmap v2427: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:41:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:42 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:41:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:42 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:41:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:42 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:41:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:42 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:41:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:42 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:41:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:42 vm01.local ceph-mon[48597]: pgmap v2427: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:41:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:42 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:41:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:42 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:41:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:42 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:41:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:42 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:41:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:42 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:41:44.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:44 vm08.local ceph-mon[50506]: pgmap v2428: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:44.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:44 vm01.local ceph-mon[48597]: pgmap v2428: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:45.376 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:41:45.376 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:41:45.404 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:41:45.404 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:41:46.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:46 vm08.local ceph-mon[50506]: pgmap v2429: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:46.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:46 vm01.local ceph-mon[48597]: pgmap v2429: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:48.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:48 vm01.local ceph-mon[48597]: pgmap v2430: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:41:48.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:48 vm08.local ceph-mon[50506]: pgmap v2430: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:41:49.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:49 vm08.local ceph-mon[50506]: pgmap v2431: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:49.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:49 vm01.local ceph-mon[48597]: pgmap v2431: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:50.405 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:41:50.406 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:41:50.432 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:41:50.432 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:41:52.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:52 vm08.local ceph-mon[50506]: pgmap v2432: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:41:52.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:52 vm01.local ceph-mon[48597]: pgmap v2432: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:41:54.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:54 vm08.local ceph-mon[50506]: pgmap v2433: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:54.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:54 vm01.local ceph-mon[48597]: pgmap v2433: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:55.433 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:41:55.434 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:41:55.460 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:41:55.460 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:41:56.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:56 vm01.local ceph-mon[48597]: pgmap v2434: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:56.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:56 vm08.local ceph-mon[50506]: pgmap v2434: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:57.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:57 vm01.local ceph-mon[48597]: pgmap v2435: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:41:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:57 vm08.local ceph-mon[50506]: pgmap v2435: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:41:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:41:59 vm08.local ceph-mon[50506]: pgmap v2436: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:41:59.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:41:59 vm01.local ceph-mon[48597]: pgmap v2436: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:00.462 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:42:00.462 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:42:00.489 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:42:00.489 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:42:02.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:02 vm08.local ceph-mon[50506]: pgmap v2437: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:42:02.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:02 vm01.local ceph-mon[48597]: pgmap v2437: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:42:04.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:04 vm08.local ceph-mon[50506]: pgmap v2438: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:04.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:04 vm01.local ceph-mon[48597]: pgmap v2438: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:05.490 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:42:05.491 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:42:05.516 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:42:05.516 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:42:06.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:06 vm08.local ceph-mon[50506]: pgmap v2439: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:06.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:06 vm01.local ceph-mon[48597]: pgmap v2439: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:08.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:08 vm01.local ceph-mon[48597]: pgmap v2440: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:08.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:08 vm08.local ceph-mon[50506]: pgmap v2440: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:09 vm08.local ceph-mon[50506]: pgmap v2441: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:42:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:09 vm01.local ceph-mon[48597]: pgmap v2441: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:42:10.517 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:42:10.518 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:42:10.545 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:42:10.545 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:42:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:42:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:42:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:42:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:42:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:11 vm08.local ceph-mon[50506]: pgmap v2442: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:11 vm01.local ceph-mon[48597]: pgmap v2442: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:14.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:14 vm08.local ceph-mon[50506]: pgmap v2443: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:42:14.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:14 vm01.local ceph-mon[48597]: pgmap v2443: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:42:15.547 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:42:15.547 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:42:15.573 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:42:15.573 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:42:16.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:16 vm08.local ceph-mon[50506]: pgmap v2444: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:42:16.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:16 vm01.local ceph-mon[48597]: pgmap v2444: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:42:18.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:18 vm08.local ceph-mon[50506]: pgmap v2445: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:18.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:18 vm01.local ceph-mon[48597]: pgmap v2445: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:19.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:19 vm08.local ceph-mon[50506]: pgmap v2446: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:19.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:19 vm01.local ceph-mon[48597]: pgmap v2446: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:20.575 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:42:20.575 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:42:20.601 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:42:20.602 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:42:22.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:22 vm08.local ceph-mon[50506]: pgmap v2447: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:42:22.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:22 vm01.local ceph-mon[48597]: pgmap v2447: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:42:24.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:24 vm01.local ceph-mon[48597]: pgmap v2448: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:24.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:24 vm08.local ceph-mon[50506]: pgmap v2448: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:25.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:25 vm08.local ceph-mon[50506]: pgmap v2449: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:25.603 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:42:25.604 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:42:25.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:25 vm01.local ceph-mon[48597]: pgmap v2449: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:25.631 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:42:25.632 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:42:28.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:28 vm08.local ceph-mon[50506]: pgmap v2450: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:42:28.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:28 vm01.local ceph-mon[48597]: pgmap v2450: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:42:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:29 vm08.local ceph-mon[50506]: pgmap v2451: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:29 vm01.local ceph-mon[48597]: pgmap v2451: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:30.633 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:42:30.634 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:42:30.661 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:42:30.662 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:42:32.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:32 vm08.local ceph-mon[50506]: pgmap v2452: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:42:32.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:32 vm01.local ceph-mon[48597]: pgmap v2452: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:42:34.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:34 vm08.local ceph-mon[50506]: pgmap v2453: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:34.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:34 vm01.local ceph-mon[48597]: pgmap v2453: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:35.663 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:42:35.664 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:42:35.690 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:42:35.690 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:42:36.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:36 vm08.local ceph-mon[50506]: pgmap v2454: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:36.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:36 vm01.local ceph-mon[48597]: pgmap v2454: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:38.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:38 vm01.local ceph-mon[48597]: pgmap v2455: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:42:38.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:38 vm08.local ceph-mon[50506]: pgmap v2455: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:42:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:39 vm08.local ceph-mon[50506]: pgmap v2456: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:39.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:39 vm01.local ceph-mon[48597]: pgmap v2456: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:40.692 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:42:40.693 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:42:40.718 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:42:40.719 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:42:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:42 vm08.local ceph-mon[50506]: pgmap v2457: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:42:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:42 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:42:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:42 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:42:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:42 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:42:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:42 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:42:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:42 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:42:42.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:42 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:42:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:42 vm01.local ceph-mon[48597]: pgmap v2457: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:42:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:42 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:42:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:42 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:42:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:42 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:42:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:42 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:42:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:42 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:42:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:42 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:42:44.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:44 vm08.local ceph-mon[50506]: pgmap v2458: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:44.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:44 vm01.local ceph-mon[48597]: pgmap v2458: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:45.720 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:42:45.721 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:42:45.748 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:42:45.748 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:42:46.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:46 vm01.local ceph-mon[48597]: pgmap v2459: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:46.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:46 vm08.local ceph-mon[50506]: pgmap v2459: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:48.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:48 vm01.local ceph-mon[48597]: pgmap v2460: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:42:48.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:48 vm08.local ceph-mon[50506]: pgmap v2460: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:42:49.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:49 vm08.local ceph-mon[50506]: pgmap v2461: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:49.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:49 vm01.local ceph-mon[48597]: pgmap v2461: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:50.749 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:42:50.750 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:42:50.777 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:42:50.778 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:42:52.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:52 vm08.local ceph-mon[50506]: pgmap v2462: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:42:52.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:52 vm01.local ceph-mon[48597]: pgmap v2462: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:42:54.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:54 vm08.local ceph-mon[50506]: pgmap v2463: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:54.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:54 vm01.local ceph-mon[48597]: pgmap v2463: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:55.779 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:42:55.780 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:42:55.805 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:42:55.806 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:42:56.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:56 vm01.local ceph-mon[48597]: pgmap v2464: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:56.478 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:56 vm08.local ceph-mon[50506]: pgmap v2464: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:58 vm01.local ceph-mon[48597]: pgmap v2465: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:42:58.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:58 vm08.local ceph-mon[50506]: pgmap v2465: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:42:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:42:59 vm08.local ceph-mon[50506]: pgmap v2466: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:42:59.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:42:59 vm01.local ceph-mon[48597]: pgmap v2466: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:00.807 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:43:00.808 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:43:00.833 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:43:00.833 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:43:02.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:02 vm01.local ceph-mon[48597]: pgmap v2467: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:43:02.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:02 vm08.local ceph-mon[50506]: pgmap v2467: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:43:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:03 vm08.local ceph-mon[50506]: pgmap v2468: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:03.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:03 vm01.local ceph-mon[48597]: pgmap v2468: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:05.835 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:43:05.835 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:43:05.860 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:43:05.861 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:43:06.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:06 vm08.local ceph-mon[50506]: pgmap v2469: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:06.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:06 vm01.local ceph-mon[48597]: pgmap v2469: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:08.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:08 vm01.local ceph-mon[48597]: pgmap v2470: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:43:08.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:08 vm08.local ceph-mon[50506]: pgmap v2470: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:43:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:09 vm08.local ceph-mon[50506]: pgmap v2471: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:09 vm01.local ceph-mon[48597]: pgmap v2471: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:43:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:43:10.862 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:43:10.863 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:43:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:43:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:43:10.888 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:43:10.888 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:43:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:11 vm08.local ceph-mon[50506]: pgmap v2472: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:43:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:11 vm01.local ceph-mon[48597]: pgmap v2472: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:43:14.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:14 vm08.local ceph-mon[50506]: pgmap v2473: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:14.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:14 vm01.local ceph-mon[48597]: pgmap v2473: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:15.890 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:43:15.890 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:43:15.916 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:43:15.916 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:43:16.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:16 vm01.local ceph-mon[48597]: pgmap v2474: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:16.478 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:16 vm08.local ceph-mon[50506]: pgmap v2474: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:18.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:18 vm01.local ceph-mon[48597]: pgmap v2475: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:43:18.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:18 vm08.local ceph-mon[50506]: pgmap v2475: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:43:19.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:19 vm08.local ceph-mon[50506]: pgmap v2476: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:19.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:19 vm01.local ceph-mon[48597]: pgmap v2476: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:20.917 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:43:20.918 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:43:20.944 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:43:20.944 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:43:22.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:22 vm01.local ceph-mon[48597]: pgmap v2477: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:43:22.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:22 vm08.local ceph-mon[50506]: pgmap v2477: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:43:24.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:24 vm01.local ceph-mon[48597]: pgmap v2478: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:24.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:24 vm08.local ceph-mon[50506]: pgmap v2478: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:25.945 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:43:25.946 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:43:25.973 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:43:25.973 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:43:26.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:26 vm01.local ceph-mon[48597]: pgmap v2479: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:26.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:26 vm08.local ceph-mon[50506]: pgmap v2479: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:27 vm08.local ceph-mon[50506]: pgmap v2480: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:43:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:27 vm01.local ceph-mon[48597]: pgmap v2480: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:43:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:29 vm08.local ceph-mon[50506]: pgmap v2481: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:29.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:29 vm01.local ceph-mon[48597]: pgmap v2481: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:30.974 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:43:30.975 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:43:31.000 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:43:31.001 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:43:32.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:32 vm01.local ceph-mon[48597]: pgmap v2482: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:43:32.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:32 vm08.local ceph-mon[50506]: pgmap v2482: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:43:34.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:34 vm01.local ceph-mon[48597]: pgmap v2483: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:34.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:34 vm08.local ceph-mon[50506]: pgmap v2483: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:36.002 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:43:36.003 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:43:36.029 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:43:36.029 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:43:36.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:36 vm01.local ceph-mon[48597]: pgmap v2484: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:36.478 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:36 vm08.local ceph-mon[50506]: pgmap v2484: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:38.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:38 vm01.local ceph-mon[48597]: pgmap v2485: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:43:38.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:38 vm08.local ceph-mon[50506]: pgmap v2485: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:43:39.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:39 vm08.local ceph-mon[50506]: pgmap v2486: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:39.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:39 vm01.local ceph-mon[48597]: pgmap v2486: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:41.030 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:43:41.031 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:43:41.057 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:43:41.057 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:43:42.239 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:42 vm08.local ceph-mon[50506]: pgmap v2487: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:43:42.239 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:42 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:43:42.239 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:42 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:43:42.239 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:42 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:43:42.239 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:42 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:43:42.239 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:42 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:43:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:42 vm01.local ceph-mon[48597]: pgmap v2487: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:43:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:42 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:43:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:42 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:43:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:42 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:43:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:42 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:43:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:42 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:43:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:43 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:43:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:43 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:43:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:43 vm08.local ceph-mon[50506]: pgmap v2488: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:43.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:43 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:43:43.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:43 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:43:43.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:43 vm01.local ceph-mon[48597]: pgmap v2488: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:46.058 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:43:46.059 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:43:46.084 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:43:46.085 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:43:46.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:46 vm01.local ceph-mon[48597]: pgmap v2489: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:46.478 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:46 vm08.local ceph-mon[50506]: pgmap v2489: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:48.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:48 vm01.local ceph-mon[48597]: pgmap v2490: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:43:48.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:48 vm08.local ceph-mon[50506]: pgmap v2490: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:43:50.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:49 vm08.local ceph-mon[50506]: pgmap v2491: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:50.117 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:49 vm01.local ceph-mon[48597]: pgmap v2491: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:51.086 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:43:51.087 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:43:51.118 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:43:51.119 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:43:52.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:52 vm01.local ceph-mon[48597]: pgmap v2492: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:43:52.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:52 vm08.local ceph-mon[50506]: pgmap v2492: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:43:54.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:54 vm01.local ceph-mon[48597]: pgmap v2493: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:54.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:54 vm08.local ceph-mon[50506]: pgmap v2493: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:56.120 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:43:56.121 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:43:56.146 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:43:56.146 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:43:56.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:56 vm01.local ceph-mon[48597]: pgmap v2494: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:56.478 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:56 vm08.local ceph-mon[50506]: pgmap v2494: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:43:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:57 vm08.local ceph-mon[50506]: pgmap v2495: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:43:57.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:57 vm01.local ceph-mon[48597]: pgmap v2495: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:44:00.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:43:59 vm08.local ceph-mon[50506]: pgmap v2496: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:00.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:43:59 vm01.local ceph-mon[48597]: pgmap v2496: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:01.147 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:44:01.148 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:44:01.173 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:44:01.174 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:44:02.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:02 vm08.local ceph-mon[50506]: pgmap v2497: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:44:02.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:02 vm01.local ceph-mon[48597]: pgmap v2497: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:44:04.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:04 vm01.local ceph-mon[48597]: pgmap v2498: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:04.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:04 vm08.local ceph-mon[50506]: pgmap v2498: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:06.175 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:44:06.175 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:44:06.201 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:44:06.201 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:44:06.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:06 vm01.local ceph-mon[48597]: pgmap v2499: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:06.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:06 vm08.local ceph-mon[50506]: pgmap v2499: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:07 vm08.local ceph-mon[50506]: pgmap v2500: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:44:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:07 vm01.local ceph-mon[48597]: pgmap v2500: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:44:10.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:09 vm08.local ceph-mon[50506]: pgmap v2501: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:10.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:09 vm01.local ceph-mon[48597]: pgmap v2501: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:44:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:44:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:44:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:44:11.202 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:44:11.203 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:44:11.229 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:44:11.229 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:44:12.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:11 vm08.local ceph-mon[50506]: pgmap v2502: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:44:12.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:11 vm01.local ceph-mon[48597]: pgmap v2502: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:44:14.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:14 vm01.local ceph-mon[48597]: pgmap v2503: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:14.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:14 vm08.local ceph-mon[50506]: pgmap v2503: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:16.230 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:44:16.231 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:44:16.259 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:44:16.259 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:44:16.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:16 vm01.local ceph-mon[48597]: pgmap v2504: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:16.396 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:16 vm08.local ceph-mon[50506]: pgmap v2504: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:18.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:18 vm01.local ceph-mon[48597]: pgmap v2505: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:44:18.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:18 vm08.local ceph-mon[50506]: pgmap v2505: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:44:19.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:19 vm01.local ceph-mon[48597]: pgmap v2506: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:19 vm08.local ceph-mon[50506]: pgmap v2506: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:21.260 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:44:21.261 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:44:21.286 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:44:21.287 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:44:22.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:22 vm01.local ceph-mon[48597]: pgmap v2507: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:44:22.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:22 vm08.local ceph-mon[50506]: pgmap v2507: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:44:24.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:24 vm01.local ceph-mon[48597]: pgmap v2508: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:24.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:24 vm08.local ceph-mon[50506]: pgmap v2508: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:26.288 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:44:26.288 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:44:26.314 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:44:26.314 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:44:26.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:26 vm01.local ceph-mon[48597]: pgmap v2509: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:26.478 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:26 vm08.local ceph-mon[50506]: pgmap v2509: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:27 vm08.local ceph-mon[50506]: pgmap v2510: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:44:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:27 vm01.local ceph-mon[48597]: pgmap v2510: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:44:29.917 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:29 vm01.local ceph-mon[48597]: pgmap v2511: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:29 vm08.local ceph-mon[50506]: pgmap v2511: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:31.316 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:44:31.316 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:44:31.342 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:44:31.343 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:44:32.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:32 vm01.local ceph-mon[48597]: pgmap v2512: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:44:32.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:32 vm08.local ceph-mon[50506]: pgmap v2512: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:44:34.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:34 vm01.local ceph-mon[48597]: pgmap v2513: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:34.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:34 vm08.local ceph-mon[50506]: pgmap v2513: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:36.344 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:44:36.344 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:44:36.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:36 vm01.local ceph-mon[48597]: pgmap v2514: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:36.371 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:44:36.372 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:44:36.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:36 vm08.local ceph-mon[50506]: pgmap v2514: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:37 vm08.local ceph-mon[50506]: pgmap v2515: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:44:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:37 vm01.local ceph-mon[48597]: pgmap v2515: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:44:40.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:39 vm08.local ceph-mon[50506]: pgmap v2516: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:39 vm01.local ceph-mon[48597]: pgmap v2516: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:41.373 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:44:41.374 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:44:41.400 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:44:41.401 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:44:42.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:42 vm01.local ceph-mon[48597]: pgmap v2517: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:44:42.564 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:42 vm08.local ceph-mon[50506]: pgmap v2517: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:44:43.123 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:43 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:44:43.124 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:43 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:44:43.124 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:43 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:44:43.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:43 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:44:43.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:43 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:44:43.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:43 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:44:44.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:44 vm01.local ceph-mon[48597]: pgmap v2518: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:44.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:44 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:44:44.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:44 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:44:44.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:44 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:44:44.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:44 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:44:44.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:44 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:44:44.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:44 vm08.local ceph-mon[50506]: pgmap v2518: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:44.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:44 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:44:44.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:44 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:44:44.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:44 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:44:44.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:44 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:44:44.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:44 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:44:45.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:45 vm08.local ceph-mon[50506]: pgmap v2519: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:45.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:45 vm01.local ceph-mon[48597]: pgmap v2519: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:46.403 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:44:46.403 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:44:46.430 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:44:46.431 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:44:48.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:48 vm01.local ceph-mon[48597]: pgmap v2520: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:44:48.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:48 vm08.local ceph-mon[50506]: pgmap v2520: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:44:49.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:49 vm01.local ceph-mon[48597]: pgmap v2521: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:50.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:49 vm08.local ceph-mon[50506]: pgmap v2521: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:51 vm01.local ceph-mon[48597]: pgmap v2522: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:44:51.432 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:44:51.433 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:44:51.500 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:44:51.501 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:44:51.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:51 vm08.local ceph-mon[50506]: pgmap v2522: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:44:54.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:54 vm01.local ceph-mon[48597]: pgmap v2523: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:54.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:54 vm08.local ceph-mon[50506]: pgmap v2523: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:56.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:56 vm08.local ceph-mon[50506]: pgmap v2524: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:56.502 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:44:56.503 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:44:56.528 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:44:56.529 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:44:56.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:56 vm01.local ceph-mon[48597]: pgmap v2524: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:44:57.237 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:57 vm08.local ceph-mon[50506]: pgmap v2525: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:44:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:57 vm01.local ceph-mon[48597]: pgmap v2525: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:45:00.038 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:44:59 vm01.local ceph-mon[48597]: pgmap v2526: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:00.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:44:59 vm08.local ceph-mon[50506]: pgmap v2526: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:01.530 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:45:01.530 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:45:01.557 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:45:01.557 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:45:02.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:02 vm01.local ceph-mon[48597]: pgmap v2527: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:45:02.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:02 vm08.local ceph-mon[50506]: pgmap v2527: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:45:04.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:04 vm01.local ceph-mon[48597]: pgmap v2528: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:04.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:04 vm08.local ceph-mon[50506]: pgmap v2528: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:05.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:05 vm08.local ceph-mon[50506]: pgmap v2529: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:05.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:05 vm01.local ceph-mon[48597]: pgmap v2529: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:06.559 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:45:06.559 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:45:06.585 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:45:06.586 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:45:08.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:08 vm01.local ceph-mon[48597]: pgmap v2530: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:45:08.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:08 vm08.local ceph-mon[50506]: pgmap v2530: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:45:10.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:09 vm08.local ceph-mon[50506]: pgmap v2531: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:10.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:09 vm01.local ceph-mon[48597]: pgmap v2531: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:45:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:45:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:45:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:45:11.587 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:45:11.587 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:45:11.613 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:45:11.614 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:45:12.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:11 vm08.local ceph-mon[50506]: pgmap v2532: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:45:12.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:11 vm01.local ceph-mon[48597]: pgmap v2532: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:45:14.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:14 vm01.local ceph-mon[48597]: pgmap v2533: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:14.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:14 vm08.local ceph-mon[50506]: pgmap v2533: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:16.396 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:16 vm08.local ceph-mon[50506]: pgmap v2534: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:16.615 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:45:16.616 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:45:16.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:16 vm01.local ceph-mon[48597]: pgmap v2534: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:16.643 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:45:16.644 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:45:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:17 vm08.local ceph-mon[50506]: pgmap v2535: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:45:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:17 vm01.local ceph-mon[48597]: pgmap v2535: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:45:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:19 vm08.local ceph-mon[50506]: pgmap v2536: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:19 vm01.local ceph-mon[48597]: pgmap v2536: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:21.645 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:45:21.646 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:45:21.674 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:45:21.674 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:45:22.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:22 vm08.local ceph-mon[50506]: pgmap v2537: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:45:22.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:22 vm01.local ceph-mon[48597]: pgmap v2537: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:45:24.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:24 vm01.local ceph-mon[48597]: pgmap v2538: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:24.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:24 vm08.local ceph-mon[50506]: pgmap v2538: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:25.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:25 vm08.local ceph-mon[50506]: pgmap v2539: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:25.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:25 vm01.local ceph-mon[48597]: pgmap v2539: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:26.676 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:45:26.676 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:45:26.702 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:45:26.702 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:45:28.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:28 vm08.local ceph-mon[50506]: pgmap v2540: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:45:28.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:28 vm01.local ceph-mon[48597]: pgmap v2540: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:45:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:29 vm08.local ceph-mon[50506]: pgmap v2541: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:29 vm01.local ceph-mon[48597]: pgmap v2541: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:31.704 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:45:31.704 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:45:31.735 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:45:31.735 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:45:32.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:32 vm08.local ceph-mon[50506]: pgmap v2542: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:45:32.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:32 vm01.local ceph-mon[48597]: pgmap v2542: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:45:34.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:34 vm01.local ceph-mon[48597]: pgmap v2543: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:34.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:34 vm08.local ceph-mon[50506]: pgmap v2543: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:35.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:35 vm08.local ceph-mon[50506]: pgmap v2544: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:35.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:35 vm01.local ceph-mon[48597]: pgmap v2544: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:36.737 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:45:36.737 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:45:36.766 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:45:36.766 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:45:38.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:38 vm08.local ceph-mon[50506]: pgmap v2545: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:45:38.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:38 vm01.local ceph-mon[48597]: pgmap v2545: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:45:40.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:39 vm08.local ceph-mon[50506]: pgmap v2546: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:39 vm01.local ceph-mon[48597]: pgmap v2546: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:41.768 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:45:41.768 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:45:41.795 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:45:41.796 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:45:42.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:42 vm08.local ceph-mon[50506]: pgmap v2547: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:45:42.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:42 vm01.local ceph-mon[48597]: pgmap v2547: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:45:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:43 vm08.local ceph-mon[50506]: pgmap v2548: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:43.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:43 vm01.local ceph-mon[48597]: pgmap v2548: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:45.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:44 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:45:45.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:44 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:45:45.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:44 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:45:45.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:44 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:45:45.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:44 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:45:45.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:44 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:45:45.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:44 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:45:45.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:44 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:45:45.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:44 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:45:45.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:44 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:45:46.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:45 vm08.local ceph-mon[50506]: pgmap v2549: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:46.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:45 vm01.local ceph-mon[48597]: pgmap v2549: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:46.798 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:45:46.798 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:45:46.829 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:45:46.829 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:45:48.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:48 vm08.local ceph-mon[50506]: pgmap v2550: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:45:48.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:48 vm01.local ceph-mon[48597]: pgmap v2550: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:45:50.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:49 vm08.local ceph-mon[50506]: pgmap v2551: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:50.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:49 vm01.local ceph-mon[48597]: pgmap v2551: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:51.831 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:45:51.831 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:45:51.857 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:45:51.857 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:45:52.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:52 vm08.local ceph-mon[50506]: pgmap v2552: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:45:52.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:52 vm01.local ceph-mon[48597]: pgmap v2552: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:45:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:53 vm08.local ceph-mon[50506]: pgmap v2553: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:53.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:53 vm01.local ceph-mon[48597]: pgmap v2553: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:56.478 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:56 vm08.local ceph-mon[50506]: pgmap v2554: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:56.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:56 vm01.local ceph-mon[48597]: pgmap v2554: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:45:56.859 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:45:56.860 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:45:56.915 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:45:56.915 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:45:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:57 vm08.local ceph-mon[50506]: pgmap v2555: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:45:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:57 vm01.local ceph-mon[48597]: pgmap v2555: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:46:00.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:45:59 vm08.local ceph-mon[50506]: pgmap v2556: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:00.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:45:59 vm01.local ceph-mon[48597]: pgmap v2556: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:01.917 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:46:01.917 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:46:01.944 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:46:01.944 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:46:02.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:02 vm08.local ceph-mon[50506]: pgmap v2557: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:46:02.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:02 vm01.local ceph-mon[48597]: pgmap v2557: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:46:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:03 vm08.local ceph-mon[50506]: pgmap v2558: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:03.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:03 vm01.local ceph-mon[48597]: pgmap v2558: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:06.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:06 vm08.local ceph-mon[50506]: pgmap v2559: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:06.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:06 vm01.local ceph-mon[48597]: pgmap v2559: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:06.946 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:46:06.947 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:46:06.976 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:46:06.976 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:46:08.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:08 vm08.local ceph-mon[50506]: pgmap v2560: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:46:08.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:08 vm01.local ceph-mon[48597]: pgmap v2560: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:46:10.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:09 vm08.local ceph-mon[50506]: pgmap v2561: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:10.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:09 vm01.local ceph-mon[48597]: pgmap v2561: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:46:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:46:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:46:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:46:11.978 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:46:11.978 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:46:12.005 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:46:12.006 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:46:12.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:11 vm08.local ceph-mon[50506]: pgmap v2562: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:46:12.117 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:11 vm01.local ceph-mon[48597]: pgmap v2562: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:46:14.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:14 vm01.local ceph-mon[48597]: pgmap v2563: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:14.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:14 vm08.local ceph-mon[50506]: pgmap v2563: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:16.396 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:16 vm08.local ceph-mon[50506]: pgmap v2564: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:16.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:16 vm01.local ceph-mon[48597]: pgmap v2564: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:17.007 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:46:17.008 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:46:17.039 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:46:17.040 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:46:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:17 vm08.local ceph-mon[50506]: pgmap v2565: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:46:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:17 vm01.local ceph-mon[48597]: pgmap v2565: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:46:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:19 vm08.local ceph-mon[50506]: pgmap v2566: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:20.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:19 vm01.local ceph-mon[48597]: pgmap v2566: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:22.040 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:46:22.041 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:46:22.067 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:46:22.068 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:46:22.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:22 vm08.local ceph-mon[50506]: pgmap v2567: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:46:22.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:22 vm01.local ceph-mon[48597]: pgmap v2567: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:46:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:23 vm08.local ceph-mon[50506]: pgmap v2568: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:23.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:23 vm01.local ceph-mon[48597]: pgmap v2568: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:26.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:26 vm08.local ceph-mon[50506]: pgmap v2569: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:26.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:26 vm01.local ceph-mon[48597]: pgmap v2569: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:27.069 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:46:27.070 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:46:27.099 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:46:27.100 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:46:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:27 vm08.local ceph-mon[50506]: pgmap v2570: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:46:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:27 vm01.local ceph-mon[48597]: pgmap v2570: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:46:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:29 vm08.local ceph-mon[50506]: pgmap v2571: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:29 vm01.local ceph-mon[48597]: pgmap v2571: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:32.101 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:46:32.102 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:46:32.128 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:46:32.128 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:46:32.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:32 vm08.local ceph-mon[50506]: pgmap v2572: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:46:32.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:32 vm01.local ceph-mon[48597]: pgmap v2572: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:46:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:33 vm08.local ceph-mon[50506]: pgmap v2573: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:33.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:33 vm01.local ceph-mon[48597]: pgmap v2573: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:36.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:36 vm08.local ceph-mon[50506]: pgmap v2574: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:36.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:36 vm01.local ceph-mon[48597]: pgmap v2574: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:37.130 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:46:37.130 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:46:37.159 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:46:37.160 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:46:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:37 vm08.local ceph-mon[50506]: pgmap v2575: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:46:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:37 vm01.local ceph-mon[48597]: pgmap v2575: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:46:40.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:39 vm08.local ceph-mon[50506]: pgmap v2576: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:39 vm01.local ceph-mon[48597]: pgmap v2576: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:42.161 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:46:42.162 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:46:42.189 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:46:42.190 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:46:42.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:42 vm08.local ceph-mon[50506]: pgmap v2577: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:46:42.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:42 vm01.local ceph-mon[48597]: pgmap v2577: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:46:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:43 vm08.local ceph-mon[50506]: pgmap v2578: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:43.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:43 vm01.local ceph-mon[48597]: pgmap v2578: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:45.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:44 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:46:45.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:44 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:46:45.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:44 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:46:45.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:44 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:46:45.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:44 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:46:45.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:44 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:46:45.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:44 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:46:45.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:44 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:46:45.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:44 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:46:45.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:44 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:46:46.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:45 vm08.local ceph-mon[50506]: pgmap v2579: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:46.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:45 vm01.local ceph-mon[48597]: pgmap v2579: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:47.191 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:46:47.192 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:46:47.218 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:46:47.219 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:46:47.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:47 vm08.local ceph-mon[50506]: pgmap v2580: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:46:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:47 vm01.local ceph-mon[48597]: pgmap v2580: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:46:50.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:49 vm08.local ceph-mon[50506]: pgmap v2581: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:50.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:49 vm01.local ceph-mon[48597]: pgmap v2581: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:51 vm01.local ceph-mon[48597]: pgmap v2582: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:46:51.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:51 vm08.local ceph-mon[50506]: pgmap v2582: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:46:52.220 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:46:52.221 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:46:52.247 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:46:52.248 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:46:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:53 vm08.local ceph-mon[50506]: pgmap v2583: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:53.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:53 vm01.local ceph-mon[48597]: pgmap v2583: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:56.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:56 vm08.local ceph-mon[50506]: pgmap v2584: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:56.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:56 vm01.local ceph-mon[48597]: pgmap v2584: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:46:57.249 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:46:57.250 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:46:57.276 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:46:57.276 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:46:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:57 vm08.local ceph-mon[50506]: pgmap v2585: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:46:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:57 vm01.local ceph-mon[48597]: pgmap v2585: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:47:00.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:46:59 vm08.local ceph-mon[50506]: pgmap v2586: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:00.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:46:59 vm01.local ceph-mon[48597]: pgmap v2586: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:02.277 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:47:02.278 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:47:02.305 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:47:02.306 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:47:02.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:02 vm08.local ceph-mon[50506]: pgmap v2587: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:47:02.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:02 vm01.local ceph-mon[48597]: pgmap v2587: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:47:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:03 vm08.local ceph-mon[50506]: pgmap v2588: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:03.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:03 vm01.local ceph-mon[48597]: pgmap v2588: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:06.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:06 vm08.local ceph-mon[50506]: pgmap v2589: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:06.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:06 vm01.local ceph-mon[48597]: pgmap v2589: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:07.307 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:47:07.308 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:47:07.333 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:47:07.334 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:47:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:07 vm08.local ceph-mon[50506]: pgmap v2590: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:47:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:07 vm01.local ceph-mon[48597]: pgmap v2590: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:47:10.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:09 vm08.local ceph-mon[50506]: pgmap v2591: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:10.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:09 vm01.local ceph-mon[48597]: pgmap v2591: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:10.992 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:47:10.992 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:47:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:47:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:47:12.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:11 vm08.local ceph-mon[50506]: pgmap v2592: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:47:12.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:11 vm01.local ceph-mon[48597]: pgmap v2592: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:47:12.335 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:47:12.336 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:47:12.364 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:47:12.364 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:47:14.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:14 vm01.local ceph-mon[48597]: pgmap v2593: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:14.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:14 vm08.local ceph-mon[50506]: pgmap v2593: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:15.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:15 vm08.local ceph-mon[50506]: pgmap v2594: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:15.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:15 vm01.local ceph-mon[48597]: pgmap v2594: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:17.366 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:47:17.367 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:47:17.393 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:47:17.393 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:47:18.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:18 vm08.local ceph-mon[50506]: pgmap v2595: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:47:18.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:18 vm01.local ceph-mon[48597]: pgmap v2595: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:47:20.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:19 vm08.local ceph-mon[50506]: pgmap v2596: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:20.117 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:19 vm01.local ceph-mon[48597]: pgmap v2596: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:22.395 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:47:22.395 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:47:22.422 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:47:22.422 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:47:22.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:22 vm08.local ceph-mon[50506]: pgmap v2597: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:47:22.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:22 vm01.local ceph-mon[48597]: pgmap v2597: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:47:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:23 vm08.local ceph-mon[50506]: pgmap v2598: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:23.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:23 vm01.local ceph-mon[48597]: pgmap v2598: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:26.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:26 vm08.local ceph-mon[50506]: pgmap v2599: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:26.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:26 vm01.local ceph-mon[48597]: pgmap v2599: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:27.424 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:47:27.424 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:47:27.453 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:47:27.454 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:47:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:27 vm08.local ceph-mon[50506]: pgmap v2600: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:47:27.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:27 vm01.local ceph-mon[48597]: pgmap v2600: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:47:30.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:29 vm08.local ceph-mon[50506]: pgmap v2601: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:30.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:29 vm01.local ceph-mon[48597]: pgmap v2601: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:32.455 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:47:32.456 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:47:32.482 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:47:32.483 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:47:32.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:32 vm08.local ceph-mon[50506]: pgmap v2602: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:47:32.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:32 vm01.local ceph-mon[48597]: pgmap v2602: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:47:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:33 vm08.local ceph-mon[50506]: pgmap v2603: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:33.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:33 vm01.local ceph-mon[48597]: pgmap v2603: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:36.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:36 vm08.local ceph-mon[50506]: pgmap v2604: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:36.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:36 vm01.local ceph-mon[48597]: pgmap v2604: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:37.484 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:47:37.485 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:47:37.511 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:47:37.512 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:47:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:37 vm08.local ceph-mon[50506]: pgmap v2605: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:47:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:37 vm01.local ceph-mon[48597]: pgmap v2605: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:47:40.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:39 vm08.local ceph-mon[50506]: pgmap v2606: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:40.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:39 vm01.local ceph-mon[48597]: pgmap v2606: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:41 vm01.local ceph-mon[48597]: pgmap v2607: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:47:41.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:41 vm08.local ceph-mon[50506]: pgmap v2607: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:47:42.514 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:47:42.514 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:47:42.541 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:47:42.542 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:47:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:43 vm08.local ceph-mon[50506]: pgmap v2608: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:43.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:43 vm01.local ceph-mon[48597]: pgmap v2608: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:44.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:44 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:47:44.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:44 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:47:44.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:44 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:47:44.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:44 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:47:44.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:44 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:47:44.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:44 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:47:46.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:45 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:47:46.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:45 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:47:46.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:45 vm08.local ceph-mon[50506]: pgmap v2609: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:46.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:45 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:47:46.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:45 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:47:46.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:45 vm01.local ceph-mon[48597]: pgmap v2609: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:47.543 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:47:47.544 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:47:47.956 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:47:47.956 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:47:48.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:47 vm08.local ceph-mon[50506]: pgmap v2610: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:47:48.140 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:47 vm01.local ceph-mon[48597]: pgmap v2610: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:47:49.621 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:49 vm01.local ceph-mon[48597]: pgmap v2611: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:49.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:49 vm08.local ceph-mon[50506]: pgmap v2611: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:51 vm01.local ceph-mon[48597]: pgmap v2612: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:47:51.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:51 vm08.local ceph-mon[50506]: pgmap v2612: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:47:52.958 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:47:52.959 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:47:52.991 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:47:52.991 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:47:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:53 vm08.local ceph-mon[50506]: pgmap v2613: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:53.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:53 vm01.local ceph-mon[48597]: pgmap v2613: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:55.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:55 vm08.local ceph-mon[50506]: pgmap v2614: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:55 vm01.local ceph-mon[48597]: pgmap v2614: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:47:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:47:57 vm08.local ceph-mon[50506]: pgmap v2615: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:47:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:47:57 vm01.local ceph-mon[48597]: pgmap v2615: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:47:57.993 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:47:57.994 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:47:58.326 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:47:58.327 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:48:00.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:00 vm01.local ceph-mon[48597]: pgmap v2616: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:00.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:00 vm08.local ceph-mon[50506]: pgmap v2616: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:02.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:02 vm01.local ceph-mon[48597]: pgmap v2617: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:48:02.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:02 vm08.local ceph-mon[50506]: pgmap v2617: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:48:03.328 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:48:03.328 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:48:03.355 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:48:03.355 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:48:04.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:04 vm01.local ceph-mon[48597]: pgmap v2618: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:04.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:04 vm08.local ceph-mon[50506]: pgmap v2618: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:06.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:06 vm08.local ceph-mon[50506]: pgmap v2619: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:06.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:06 vm01.local ceph-mon[48597]: pgmap v2619: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:07.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:07 vm08.local ceph-mon[50506]: pgmap v2620: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:48:07.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:07 vm01.local ceph-mon[48597]: pgmap v2620: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:48:08.357 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:48:08.357 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:48:08.382 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:48:08.383 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:48:10.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:09 vm08.local ceph-mon[50506]: pgmap v2621: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:10.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:09 vm01.local ceph-mon[48597]: pgmap v2621: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:48:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:48:11.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:48:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:48:12.343 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:12 vm08.local ceph-mon[50506]: pgmap v2622: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:48:12.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:12 vm01.local ceph-mon[48597]: pgmap v2622: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:48:13.384 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:48:13.384 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:48:13.409 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:48:13.410 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:48:13.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:13 vm08.local ceph-mon[50506]: pgmap v2623: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:13.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:13 vm01.local ceph-mon[48597]: pgmap v2623: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:15.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:15 vm08.local ceph-mon[50506]: pgmap v2624: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:15.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:15 vm01.local ceph-mon[48597]: pgmap v2624: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:17 vm08.local ceph-mon[50506]: pgmap v2625: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:48:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:17 vm01.local ceph-mon[48597]: pgmap v2625: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:48:18.411 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:48:18.412 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:48:18.436 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:48:18.437 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:48:20.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:20 vm08.local ceph-mon[50506]: pgmap v2626: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:20.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:20 vm01.local ceph-mon[48597]: pgmap v2626: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:21.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:21 vm01.local ceph-mon[48597]: pgmap v2627: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:48:21.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:21 vm08.local ceph-mon[50506]: pgmap v2627: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:48:23.438 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:48:23.439 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:48:23.465 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:48:23.466 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:48:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:23 vm08.local ceph-mon[50506]: pgmap v2628: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:23.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:23 vm01.local ceph-mon[48597]: pgmap v2628: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:25.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:25 vm08.local ceph-mon[50506]: pgmap v2629: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:25.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:25 vm01.local ceph-mon[48597]: pgmap v2629: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:27 vm08.local ceph-mon[50506]: pgmap v2630: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:48:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:27 vm01.local ceph-mon[48597]: pgmap v2630: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:48:28.467 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:48:28.468 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:48:28.494 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:48:28.495 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:48:30.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:30 vm08.local ceph-mon[50506]: pgmap v2631: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:30.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:30 vm01.local ceph-mon[48597]: pgmap v2631: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:31.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:31 vm01.local ceph-mon[48597]: pgmap v2632: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:48:31.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:31 vm08.local ceph-mon[50506]: pgmap v2632: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:48:33.496 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:48:33.497 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:48:33.525 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:48:33.526 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:48:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:33 vm08.local ceph-mon[50506]: pgmap v2633: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:33.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:33 vm01.local ceph-mon[48597]: pgmap v2633: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:35.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:35 vm08.local ceph-mon[50506]: pgmap v2634: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:35.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:35 vm01.local ceph-mon[48597]: pgmap v2634: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:37 vm08.local ceph-mon[50506]: pgmap v2635: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:48:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:37 vm01.local ceph-mon[48597]: pgmap v2635: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:48:38.527 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:48:38.528 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:48:38.555 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:48:38.556 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:48:40.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:40 vm08.local ceph-mon[50506]: pgmap v2636: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:40.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:40 vm01.local ceph-mon[48597]: pgmap v2636: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:41 vm01.local ceph-mon[48597]: pgmap v2637: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:48:41.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:41 vm08.local ceph-mon[50506]: pgmap v2637: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:48:43.557 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:48:43.557 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:48:43.584 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:48:43.585 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:48:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:43 vm08.local ceph-mon[50506]: pgmap v2638: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:43.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:43 vm01.local ceph-mon[48597]: pgmap v2638: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:45.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:45 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:48:45.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:45 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:48:45.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:45 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:48:45.117 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:45 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:48:45.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:45 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:48:45.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:45 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:48:46.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:46 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:48:46.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:46 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:48:46.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:46 vm08.local ceph-mon[50506]: pgmap v2639: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:46.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:46 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:48:46.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:46 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:48:46.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:46 vm01.local ceph-mon[48597]: pgmap v2639: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:47.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:47 vm08.local ceph-mon[50506]: pgmap v2640: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:48:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:47 vm01.local ceph-mon[48597]: pgmap v2640: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:48:48.587 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:48:48.587 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:48:48.613 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:48:48.613 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:48:50.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:50 vm08.local ceph-mon[50506]: pgmap v2641: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:50.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:50 vm01.local ceph-mon[48597]: pgmap v2641: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:51 vm01.local ceph-mon[48597]: pgmap v2642: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:48:51.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:51 vm08.local ceph-mon[50506]: pgmap v2642: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:48:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:53 vm08.local ceph-mon[50506]: pgmap v2643: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:53.615 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:48:53.615 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:48:53.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:53 vm01.local ceph-mon[48597]: pgmap v2643: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:53.642 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:48:53.643 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:48:55.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:55 vm08.local ceph-mon[50506]: pgmap v2644: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:55 vm01.local ceph-mon[48597]: pgmap v2644: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:48:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:48:57 vm08.local ceph-mon[50506]: pgmap v2645: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:48:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:48:57 vm01.local ceph-mon[48597]: pgmap v2645: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:48:58.644 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:48:58.644 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:48:58.671 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:48:58.672 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:49:00.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:00 vm08.local ceph-mon[50506]: pgmap v2646: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:00.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:00 vm01.local ceph-mon[48597]: pgmap v2646: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:01 vm08.local ceph-mon[50506]: pgmap v2647: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:49:01.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:01 vm01.local ceph-mon[48597]: pgmap v2647: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:49:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:03 vm08.local ceph-mon[50506]: pgmap v2648: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:03.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:03 vm01.local ceph-mon[48597]: pgmap v2648: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:03.673 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:49:03.674 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:49:04.073 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:49:04.074 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:49:05.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:05 vm08.local ceph-mon[50506]: pgmap v2649: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:05.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:05 vm01.local ceph-mon[48597]: pgmap v2649: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:08.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:07 vm08.local ceph-mon[50506]: pgmap v2650: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:49:08.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:07 vm01.local ceph-mon[48597]: pgmap v2650: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:49:09.075 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:49:09.076 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:49:09.355 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:49:09.397 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:49:09.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:09 vm01.local ceph-mon[48597]: pgmap v2651: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:09 vm08.local ceph-mon[50506]: pgmap v2651: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:49:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:49:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:49:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:11 vm08.local ceph-mon[50506]: pgmap v2652: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:49:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:49:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:11 vm01.local ceph-mon[48597]: pgmap v2652: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:49:13.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:13 vm08.local ceph-mon[50506]: pgmap v2653: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:13.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:13 vm01.local ceph-mon[48597]: pgmap v2653: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:14.357 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:49:14.358 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:49:14.383 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:49:14.383 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:49:16.478 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:16 vm08.local ceph-mon[50506]: pgmap v2654: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:16.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:16 vm01.local ceph-mon[48597]: pgmap v2654: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:18.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:18 vm08.local ceph-mon[50506]: pgmap v2655: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:49:18.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:18 vm01.local ceph-mon[48597]: pgmap v2655: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:49:19.384 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:49:19.385 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:49:19.411 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:49:19.412 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:49:20.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:20 vm08.local ceph-mon[50506]: pgmap v2656: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:20.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:20 vm01.local ceph-mon[48597]: pgmap v2656: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:22.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:22 vm08.local ceph-mon[50506]: pgmap v2657: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:49:22.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:22 vm01.local ceph-mon[48597]: pgmap v2657: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:49:24.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:24 vm01.local ceph-mon[48597]: pgmap v2658: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:24.413 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:49:24.414 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:49:24.440 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:49:24.440 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:49:24.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:24 vm08.local ceph-mon[50506]: pgmap v2658: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:25.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:25 vm08.local ceph-mon[50506]: pgmap v2659: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:25.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:25 vm01.local ceph-mon[48597]: pgmap v2659: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:27 vm08.local ceph-mon[50506]: pgmap v2660: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:49:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:27 vm01.local ceph-mon[48597]: pgmap v2660: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:49:29.441 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:49:29.442 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:49:29.469 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:49:29.469 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:49:30.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:30 vm08.local ceph-mon[50506]: pgmap v2661: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:30.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:30 vm01.local ceph-mon[48597]: pgmap v2661: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:32.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:32 vm08.local ceph-mon[50506]: pgmap v2662: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:49:32.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:32 vm01.local ceph-mon[48597]: pgmap v2662: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:49:34.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:34 vm01.local ceph-mon[48597]: pgmap v2663: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:34.470 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:49:34.471 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:49:34.497 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:49:34.497 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:49:34.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:34 vm08.local ceph-mon[50506]: pgmap v2663: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:35.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:35 vm08.local ceph-mon[50506]: pgmap v2664: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:35.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:35 vm01.local ceph-mon[48597]: pgmap v2664: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:37 vm08.local ceph-mon[50506]: pgmap v2665: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:49:37.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:37 vm01.local ceph-mon[48597]: pgmap v2665: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:49:39.498 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:49:39.499 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:49:39.524 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:49:39.525 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:49:40.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:40 vm08.local ceph-mon[50506]: pgmap v2666: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:40.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:40 vm01.local ceph-mon[48597]: pgmap v2666: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:41.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:41 vm08.local ceph-mon[50506]: pgmap v2667: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:49:41.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:41 vm01.local ceph-mon[48597]: pgmap v2667: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:49:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:43 vm08.local ceph-mon[50506]: pgmap v2668: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:43.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:43 vm01.local ceph-mon[48597]: pgmap v2668: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:44.526 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:49:44.526 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:49:44.552 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:49:44.552 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:49:45.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:45 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:49:45.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:45 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:49:45.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:45 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:49:45.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:45 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:49:45.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:45 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:49:45.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:45 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:49:46.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:46 vm08.local ceph-mon[50506]: pgmap v2669: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:46.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:46 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:49:46.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:46 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:49:46.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:46 vm01.local ceph-mon[48597]: pgmap v2669: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:46.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:46 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:49:46.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:46 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:49:47.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:47 vm08.local ceph-mon[50506]: pgmap v2670: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:49:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:47 vm01.local ceph-mon[48597]: pgmap v2670: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:49:49.554 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:49:49.555 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:49:49.581 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:49:49.581 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:49:50.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:50 vm08.local ceph-mon[50506]: pgmap v2671: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:50.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:50 vm01.local ceph-mon[48597]: pgmap v2671: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:52.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:52 vm08.local ceph-mon[50506]: pgmap v2672: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:49:52.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:52 vm01.local ceph-mon[48597]: pgmap v2672: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:49:54.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:54 vm01.local ceph-mon[48597]: pgmap v2673: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:54.582 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:49:54.583 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:49:54.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:54 vm08.local ceph-mon[50506]: pgmap v2673: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:54.608 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:49:54.609 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:49:55.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:55 vm08.local ceph-mon[50506]: pgmap v2674: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:55.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:55 vm01.local ceph-mon[48597]: pgmap v2674: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:49:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:49:57 vm08.local ceph-mon[50506]: pgmap v2675: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:49:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:49:57 vm01.local ceph-mon[48597]: pgmap v2675: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:49:59.610 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:49:59.610 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:49:59.636 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:49:59.636 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:50:00.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:00 vm08.local ceph-mon[50506]: pgmap v2676: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:00.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:00 vm08.local ceph-mon[50506]: overall HEALTH_OK 2026-03-10T11:50:00.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:00 vm01.local ceph-mon[48597]: pgmap v2676: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:00.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:00 vm01.local ceph-mon[48597]: overall HEALTH_OK 2026-03-10T11:50:02.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:02 vm08.local ceph-mon[50506]: pgmap v2677: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:50:02.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:02 vm01.local ceph-mon[48597]: pgmap v2677: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:50:03.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:03 vm08.local ceph-mon[50506]: pgmap v2678: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:03.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:03 vm01.local ceph-mon[48597]: pgmap v2678: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:04.638 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:50:04.638 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:50:04.664 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:50:04.664 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:50:06.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:06 vm08.local ceph-mon[50506]: pgmap v2679: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:06.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:06 vm01.local ceph-mon[48597]: pgmap v2679: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:08.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:08 vm08.local ceph-mon[50506]: pgmap v2680: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:50:08.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:08 vm01.local ceph-mon[48597]: pgmap v2680: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:50:09.665 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:50:09.666 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:50:09.691 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:50:09.692 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:50:10.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:10 vm08.local ceph-mon[50506]: pgmap v2681: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:10.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:10 vm01.local ceph-mon[48597]: pgmap v2681: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:11.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:50:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:50:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:50:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:50:12.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:12 vm08.local ceph-mon[50506]: pgmap v2682: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:50:12.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:12 vm01.local ceph-mon[48597]: pgmap v2682: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:50:14.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:14 vm01.local ceph-mon[48597]: pgmap v2683: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:14.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:14 vm08.local ceph-mon[50506]: pgmap v2683: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:14.693 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:50:14.693 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:50:14.719 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:50:14.720 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:50:15.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:15 vm08.local ceph-mon[50506]: pgmap v2684: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:15.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:15 vm01.local ceph-mon[48597]: pgmap v2684: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:17 vm08.local ceph-mon[50506]: pgmap v2685: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:50:17.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:17 vm01.local ceph-mon[48597]: pgmap v2685: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:50:19.721 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:50:19.722 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:50:19.747 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:50:19.748 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:50:20.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:20 vm08.local ceph-mon[50506]: pgmap v2686: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:20.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:20 vm01.local ceph-mon[48597]: pgmap v2686: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:22.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:22 vm08.local ceph-mon[50506]: pgmap v2687: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:50:22.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:22 vm01.local ceph-mon[48597]: pgmap v2687: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:50:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:23 vm08.local ceph-mon[50506]: pgmap v2688: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:23.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:23 vm01.local ceph-mon[48597]: pgmap v2688: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:24.749 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:50:24.749 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:50:24.776 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:50:24.777 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:50:26.478 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:26 vm08.local ceph-mon[50506]: pgmap v2689: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:26.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:26 vm01.local ceph-mon[48597]: pgmap v2689: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:27 vm08.local ceph-mon[50506]: pgmap v2690: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:50:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:27 vm01.local ceph-mon[48597]: pgmap v2690: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:50:29.778 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:50:29.779 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:50:29.807 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:50:29.808 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:50:30.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:30 vm08.local ceph-mon[50506]: pgmap v2691: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:30.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:30 vm01.local ceph-mon[48597]: pgmap v2691: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:31.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:31 vm01.local ceph-mon[48597]: pgmap v2692: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:50:31.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:31 vm08.local ceph-mon[50506]: pgmap v2692: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:50:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:33 vm08.local ceph-mon[50506]: pgmap v2693: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:33.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:33 vm01.local ceph-mon[48597]: pgmap v2693: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:34.809 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:50:34.810 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:50:34.836 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:50:34.837 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:50:36.478 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:36 vm08.local ceph-mon[50506]: pgmap v2694: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:36.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:36 vm01.local ceph-mon[48597]: pgmap v2694: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:37 vm08.local ceph-mon[50506]: pgmap v2695: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:50:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:37 vm01.local ceph-mon[48597]: pgmap v2695: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:50:39.838 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:50:39.838 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:50:39.865 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:50:39.865 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:50:40.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:40 vm08.local ceph-mon[50506]: pgmap v2696: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:40.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:40 vm01.local ceph-mon[48597]: pgmap v2696: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:41 vm01.local ceph-mon[48597]: pgmap v2697: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:50:41.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:41 vm08.local ceph-mon[50506]: pgmap v2697: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:50:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:43 vm08.local ceph-mon[50506]: pgmap v2698: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:43.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:43 vm01.local ceph-mon[48597]: pgmap v2698: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:44.867 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:50:44.867 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:50:44.892 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:50:44.893 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:50:45.491 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:45 vm01.local ceph-mon[48597]: pgmap v2699: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:45.495 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:45 vm08.local ceph-mon[50506]: pgmap v2699: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:46.480 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:46 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:50:46.480 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:46 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:50:46.480 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:46 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:50:46.480 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:46 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:50:46.480 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:46 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:50:46.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:46 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:50:46.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:46 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:50:46.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:46 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:50:46.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:46 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:50:46.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:46 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:50:47.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:47 vm08.local ceph-mon[50506]: pgmap v2700: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:50:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:47 vm01.local ceph-mon[48597]: pgmap v2700: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:50:49.894 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:50:49.894 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:50:49.920 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:50:49.921 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:50:50.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:50 vm08.local ceph-mon[50506]: pgmap v2701: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:50.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:50 vm01.local ceph-mon[48597]: pgmap v2701: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:51 vm01.local ceph-mon[48597]: pgmap v2702: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:50:51.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:51 vm08.local ceph-mon[50506]: pgmap v2702: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:50:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:53 vm08.local ceph-mon[50506]: pgmap v2703: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:53.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:53 vm01.local ceph-mon[48597]: pgmap v2703: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:54.922 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:50:54.923 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:50:54.954 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:50:54.954 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:50:55.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:55 vm08.local ceph-mon[50506]: pgmap v2704: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:55 vm01.local ceph-mon[48597]: pgmap v2704: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:57 vm08.local ceph-mon[50506]: pgmap v2705: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:50:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:57 vm01.local ceph-mon[48597]: pgmap v2705: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:50:59.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:50:59 vm01.local ceph-mon[48597]: pgmap v2706: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:50:59 vm08.local ceph-mon[50506]: pgmap v2706: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:50:59.956 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:50:59.956 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:50:59.983 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:50:59.984 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:51:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:01 vm01.local ceph-mon[48597]: pgmap v2707: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:51:01.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:01 vm08.local ceph-mon[50506]: pgmap v2707: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:51:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:03 vm08.local ceph-mon[50506]: pgmap v2708: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:03.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:03 vm01.local ceph-mon[48597]: pgmap v2708: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:04.985 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:51:04.986 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:51:05.013 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:51:05.014 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:51:05.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:05 vm08.local ceph-mon[50506]: pgmap v2709: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:05.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:05 vm01.local ceph-mon[48597]: pgmap v2709: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:07 vm08.local ceph-mon[50506]: pgmap v2710: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:51:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:07 vm01.local ceph-mon[48597]: pgmap v2710: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:51:09.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:09 vm01.local ceph-mon[48597]: pgmap v2711: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:09 vm08.local ceph-mon[50506]: pgmap v2711: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:10.015 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:51:10.016 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:51:10.041 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:51:10.042 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:51:10.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:51:10.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:51:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:51:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:11 vm01.local ceph-mon[48597]: pgmap v2712: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:51:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:51:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:11 vm08.local ceph-mon[50506]: pgmap v2712: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:51:13.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:13 vm08.local ceph-mon[50506]: pgmap v2713: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:13.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:13 vm01.local ceph-mon[48597]: pgmap v2713: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:15.043 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:51:15.044 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:51:15.069 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:51:15.070 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:51:15.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:15 vm08.local ceph-mon[50506]: pgmap v2714: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:15.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:15 vm01.local ceph-mon[48597]: pgmap v2714: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:17 vm08.local ceph-mon[50506]: pgmap v2715: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:51:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:17 vm01.local ceph-mon[48597]: pgmap v2715: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:51:19.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:19 vm01.local ceph-mon[48597]: pgmap v2716: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:19.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:19 vm08.local ceph-mon[50506]: pgmap v2716: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:20.071 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:51:20.072 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:51:20.099 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:51:20.099 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:51:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:21 vm01.local ceph-mon[48597]: pgmap v2717: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:51:21.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:21 vm08.local ceph-mon[50506]: pgmap v2717: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:51:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:23 vm08.local ceph-mon[50506]: pgmap v2718: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:23.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:23 vm01.local ceph-mon[48597]: pgmap v2718: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:25.101 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:51:25.101 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:51:25.127 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:51:25.128 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:51:25.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:25 vm08.local ceph-mon[50506]: pgmap v2719: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:25.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:25 vm01.local ceph-mon[48597]: pgmap v2719: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:27 vm08.local ceph-mon[50506]: pgmap v2720: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:51:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:27 vm01.local ceph-mon[48597]: pgmap v2720: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:51:29.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:29 vm01.local ceph-mon[48597]: pgmap v2721: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:29 vm08.local ceph-mon[50506]: pgmap v2721: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:30.129 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:51:30.129 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:51:30.155 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:51:30.155 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:51:31.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:31 vm01.local ceph-mon[48597]: pgmap v2722: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:51:31.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:31 vm08.local ceph-mon[50506]: pgmap v2722: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:51:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:33 vm08.local ceph-mon[50506]: pgmap v2723: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:33.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:33 vm01.local ceph-mon[48597]: pgmap v2723: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:35.156 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:51:35.157 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:51:35.184 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:51:35.184 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:51:35.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:35 vm08.local ceph-mon[50506]: pgmap v2724: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:35.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:35 vm01.local ceph-mon[48597]: pgmap v2724: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:37 vm08.local ceph-mon[50506]: pgmap v2725: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:51:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:37 vm01.local ceph-mon[48597]: pgmap v2725: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:51:39.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:39 vm01.local ceph-mon[48597]: pgmap v2726: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:39.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:39 vm08.local ceph-mon[50506]: pgmap v2726: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:40.186 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:51:40.186 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:51:40.212 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:51:40.212 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:51:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:41 vm01.local ceph-mon[48597]: pgmap v2727: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:51:41.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:41 vm08.local ceph-mon[50506]: pgmap v2727: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:51:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:43 vm08.local ceph-mon[50506]: pgmap v2728: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:43.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:43 vm01.local ceph-mon[48597]: pgmap v2728: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:45.213 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:51:45.214 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:51:45.240 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:51:45.240 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:51:45.532 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:45 vm01.local ceph-mon[48597]: pgmap v2729: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:45.535 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:45 vm08.local ceph-mon[50506]: pgmap v2729: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:46.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:46 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:51:46.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:46 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:51:46.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:46 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:51:46.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:46 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:51:46.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:46 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:51:46.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:46 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:51:46.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:46 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:51:46.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:46 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:51:46.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:46 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:51:46.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:46 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:51:47.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:47 vm08.local ceph-mon[50506]: pgmap v2730: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:51:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:47 vm01.local ceph-mon[48597]: pgmap v2730: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:51:49.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:49 vm01.local ceph-mon[48597]: pgmap v2731: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:49.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:49 vm08.local ceph-mon[50506]: pgmap v2731: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:50.241 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:51:50.242 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:51:50.267 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:51:50.268 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:51:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:51 vm01.local ceph-mon[48597]: pgmap v2732: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:51:51.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:51 vm08.local ceph-mon[50506]: pgmap v2732: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:51:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:53 vm08.local ceph-mon[50506]: pgmap v2733: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:53.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:53 vm01.local ceph-mon[48597]: pgmap v2733: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:55.269 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:51:55.269 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:51:55.295 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:51:55.296 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:51:55.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:55 vm08.local ceph-mon[50506]: pgmap v2734: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:55 vm01.local ceph-mon[48597]: pgmap v2734: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:57 vm08.local ceph-mon[50506]: pgmap v2735: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:51:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:57 vm01.local ceph-mon[48597]: pgmap v2735: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:51:59.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:51:59 vm01.local ceph-mon[48597]: pgmap v2736: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:51:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:51:59 vm08.local ceph-mon[50506]: pgmap v2736: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:00.297 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:52:00.297 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:52:00.322 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:52:00.323 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:52:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:01 vm01.local ceph-mon[48597]: pgmap v2737: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:52:01.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:01 vm08.local ceph-mon[50506]: pgmap v2737: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:52:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:03 vm08.local ceph-mon[50506]: pgmap v2738: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:03.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:03 vm01.local ceph-mon[48597]: pgmap v2738: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:05.324 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:52:05.324 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:52:05.351 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:52:05.351 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:52:05.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:05 vm08.local ceph-mon[50506]: pgmap v2739: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:05.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:05 vm01.local ceph-mon[48597]: pgmap v2739: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:07 vm08.local ceph-mon[50506]: pgmap v2740: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:52:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:07 vm01.local ceph-mon[48597]: pgmap v2740: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:52:09.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:09 vm01.local ceph-mon[48597]: pgmap v2741: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:09 vm08.local ceph-mon[50506]: pgmap v2741: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:10.353 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:52:10.353 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:52:10.379 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:52:10.379 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:52:10.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:52:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:52:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:52:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:11 vm01.local ceph-mon[48597]: pgmap v2742: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:52:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:52:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:11 vm08.local ceph-mon[50506]: pgmap v2742: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:52:13.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:13 vm08.local ceph-mon[50506]: pgmap v2743: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:13.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:13 vm01.local ceph-mon[48597]: pgmap v2743: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:15.380 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:52:15.381 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:52:15.407 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:52:15.407 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:52:15.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:15 vm08.local ceph-mon[50506]: pgmap v2744: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:15.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:15 vm01.local ceph-mon[48597]: pgmap v2744: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:17 vm08.local ceph-mon[50506]: pgmap v2745: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:52:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:17 vm01.local ceph-mon[48597]: pgmap v2745: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:52:19.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:19 vm01.local ceph-mon[48597]: pgmap v2746: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:19.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:19 vm08.local ceph-mon[50506]: pgmap v2746: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:20.409 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:52:20.409 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:52:20.443 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:52:20.443 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:52:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:21 vm01.local ceph-mon[48597]: pgmap v2747: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:52:21.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:21 vm08.local ceph-mon[50506]: pgmap v2747: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:52:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:23 vm08.local ceph-mon[50506]: pgmap v2748: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:23.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:23 vm01.local ceph-mon[48597]: pgmap v2748: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:25.444 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:52:25.445 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:52:25.471 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:52:25.471 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:52:25.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:25 vm08.local ceph-mon[50506]: pgmap v2749: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:25.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:25 vm01.local ceph-mon[48597]: pgmap v2749: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:27 vm08.local ceph-mon[50506]: pgmap v2750: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:52:27.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:27 vm01.local ceph-mon[48597]: pgmap v2750: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:52:29.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:29 vm01.local ceph-mon[48597]: pgmap v2751: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:29 vm08.local ceph-mon[50506]: pgmap v2751: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:30.473 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:52:30.473 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:52:30.500 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:52:30.501 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:52:31.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:31 vm01.local ceph-mon[48597]: pgmap v2752: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:52:31.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:31 vm08.local ceph-mon[50506]: pgmap v2752: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:52:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:33 vm08.local ceph-mon[50506]: pgmap v2753: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:33.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:33 vm01.local ceph-mon[48597]: pgmap v2753: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:35.502 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:52:35.503 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:52:35.528 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:52:35.529 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:52:35.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:35 vm08.local ceph-mon[50506]: pgmap v2754: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:35.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:35 vm01.local ceph-mon[48597]: pgmap v2754: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:37 vm08.local ceph-mon[50506]: pgmap v2755: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:52:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:37 vm01.local ceph-mon[48597]: pgmap v2755: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:52:39.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:39 vm01.local ceph-mon[48597]: pgmap v2756: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:39 vm08.local ceph-mon[50506]: pgmap v2756: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:40.530 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:52:40.530 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:52:40.557 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:52:40.558 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:52:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:41 vm01.local ceph-mon[48597]: pgmap v2757: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:41.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:41 vm08.local ceph-mon[50506]: pgmap v2757: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:43 vm08.local ceph-mon[50506]: pgmap v2758: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:52:43.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:43 vm01.local ceph-mon[48597]: pgmap v2758: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:52:45.559 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:52:45.560 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:52:45.586 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:52:45.586 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:52:45.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:45 vm08.local ceph-mon[50506]: pgmap v2759: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:52:45.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:45 vm01.local ceph-mon[48597]: pgmap v2759: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:52:46.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:46 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:52:46.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:46 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:52:46.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:46 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:52:46.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:46 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:52:46.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:46 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:52:46.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:46 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:52:46.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:46 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:52:46.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:46 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:52:46.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:46 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:52:46.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:46 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:52:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:47 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:52:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:47 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:52:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:47 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:52:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:47 vm01.local ceph-mon[48597]: pgmap v2760: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:47.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:47 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:52:47.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:47 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:52:47.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:47 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:52:47.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:47 vm08.local ceph-mon[50506]: pgmap v2760: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:49.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:49 vm01.local ceph-mon[48597]: pgmap v2761: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:52:49.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:49 vm08.local ceph-mon[50506]: pgmap v2761: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T11:52:50.588 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:52:50.588 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:52:50.614 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:52:50.614 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:52:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:51 vm01.local ceph-mon[48597]: pgmap v2762: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:51.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:51 vm08.local ceph-mon[50506]: pgmap v2762: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:53 vm08.local ceph-mon[50506]: pgmap v2763: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:53.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:53 vm01.local ceph-mon[48597]: pgmap v2763: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:55.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:55 vm08.local ceph-mon[50506]: pgmap v2764: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:55.615 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:52:55.616 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:52:55.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:55 vm01.local ceph-mon[48597]: pgmap v2764: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:55.642 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:52:55.642 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:52:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:57 vm08.local ceph-mon[50506]: pgmap v2765: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:52:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:57 vm01.local ceph-mon[48597]: pgmap v2765: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:52:59.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:52:59 vm01.local ceph-mon[48597]: pgmap v2766: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:52:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:52:59 vm08.local ceph-mon[50506]: pgmap v2766: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:00.644 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:53:00.644 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:53:00.670 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:53:00.671 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:53:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:01 vm01.local ceph-mon[48597]: pgmap v2767: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:53:01.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:01 vm08.local ceph-mon[50506]: pgmap v2767: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:53:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:03 vm08.local ceph-mon[50506]: pgmap v2768: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:03.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:03 vm01.local ceph-mon[48597]: pgmap v2768: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:05.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:05 vm08.local ceph-mon[50506]: pgmap v2769: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:05.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:05 vm01.local ceph-mon[48597]: pgmap v2769: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:05.672 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:53:05.673 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:53:05.698 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:53:05.699 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:53:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:07 vm08.local ceph-mon[50506]: pgmap v2770: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:53:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:07 vm01.local ceph-mon[48597]: pgmap v2770: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:53:09.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:09 vm01.local ceph-mon[48597]: pgmap v2771: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:09 vm08.local ceph-mon[50506]: pgmap v2771: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:10.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:53:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:53:10.700 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:53:10.700 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:53:10.727 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:53:10.727 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:53:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:53:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:11 vm08.local ceph-mon[50506]: pgmap v2772: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:53:11.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:53:11.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:11 vm01.local ceph-mon[48597]: pgmap v2772: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:53:13.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:13 vm08.local ceph-mon[50506]: pgmap v2773: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:13.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:13 vm01.local ceph-mon[48597]: pgmap v2773: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:15.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:15 vm08.local ceph-mon[50506]: pgmap v2774: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:15.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:15 vm01.local ceph-mon[48597]: pgmap v2774: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:15.729 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:53:15.729 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:53:15.757 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:53:15.757 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:53:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:17 vm08.local ceph-mon[50506]: pgmap v2775: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:53:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:17 vm01.local ceph-mon[48597]: pgmap v2775: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:53:19.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:19 vm01.local ceph-mon[48597]: pgmap v2776: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:19.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:19 vm08.local ceph-mon[50506]: pgmap v2776: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:20.759 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:53:20.759 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:53:20.786 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:53:20.786 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:53:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:21 vm01.local ceph-mon[48597]: pgmap v2777: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:53:21.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:21 vm08.local ceph-mon[50506]: pgmap v2777: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:53:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:23 vm08.local ceph-mon[50506]: pgmap v2778: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:23.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:23 vm01.local ceph-mon[48597]: pgmap v2778: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:25.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:25 vm08.local ceph-mon[50506]: pgmap v2779: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:25.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:25 vm01.local ceph-mon[48597]: pgmap v2779: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:25.788 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:53:25.788 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:53:25.818 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:53:25.818 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:53:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:27 vm08.local ceph-mon[50506]: pgmap v2780: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:53:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:27 vm01.local ceph-mon[48597]: pgmap v2780: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:53:29.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:29 vm01.local ceph-mon[48597]: pgmap v2781: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:29 vm08.local ceph-mon[50506]: pgmap v2781: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:30.820 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:53:30.820 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:53:30.847 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:53:30.847 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:53:31.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:31 vm01.local ceph-mon[48597]: pgmap v2782: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:53:31.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:31 vm08.local ceph-mon[50506]: pgmap v2782: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:53:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:33 vm08.local ceph-mon[50506]: pgmap v2783: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:33.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:33 vm01.local ceph-mon[48597]: pgmap v2783: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:35.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:35 vm08.local ceph-mon[50506]: pgmap v2784: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:35.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:35 vm01.local ceph-mon[48597]: pgmap v2784: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:35.848 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:53:35.849 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:53:35.875 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:53:35.876 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:53:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:37 vm08.local ceph-mon[50506]: pgmap v2785: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:53:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:37 vm01.local ceph-mon[48597]: pgmap v2785: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:53:39.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:39 vm01.local ceph-mon[48597]: pgmap v2786: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:39 vm08.local ceph-mon[50506]: pgmap v2786: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:40.877 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:53:40.877 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:53:40.903 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:53:40.903 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:53:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:41 vm01.local ceph-mon[48597]: pgmap v2787: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:53:41.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:41 vm08.local ceph-mon[50506]: pgmap v2787: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:53:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:43 vm08.local ceph-mon[50506]: pgmap v2788: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:43.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:43 vm01.local ceph-mon[48597]: pgmap v2788: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:45.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:45 vm08.local ceph-mon[50506]: pgmap v2789: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:45.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:45 vm01.local ceph-mon[48597]: pgmap v2789: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:45.905 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:53:45.905 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:53:45.933 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:53:45.934 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:53:46.643 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:46 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:53:46.643 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:46 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:53:46.643 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:46 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:53:46.813 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:46 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:53:46.813 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:46 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:53:46.813 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:46 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:53:47.793 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:47 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:53:47.793 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:47 vm01.local ceph-mon[48597]: pgmap v2790: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:53:47.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:47 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:53:47.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:47 vm08.local ceph-mon[50506]: pgmap v2790: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:53:49.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:48 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:53:49.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:48 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:53:49.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:48 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:53:49.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:48 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:53:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:48 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:53:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:48 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:53:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:48 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:53:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:48 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:53:50.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:49 vm01.local ceph-mon[48597]: pgmap v2791: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:50.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:49 vm08.local ceph-mon[50506]: pgmap v2791: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:50.936 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:53:50.937 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:53:50.963 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:53:50.964 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:53:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:51 vm01.local ceph-mon[48597]: pgmap v2792: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:53:51.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:51 vm08.local ceph-mon[50506]: pgmap v2792: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:53:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:53 vm08.local ceph-mon[50506]: pgmap v2793: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:53.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:53 vm01.local ceph-mon[48597]: pgmap v2793: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:55.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:55 vm08.local ceph-mon[50506]: pgmap v2794: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:55 vm01.local ceph-mon[48597]: pgmap v2794: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:55.965 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:53:55.965 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:53:55.991 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:53:55.991 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:53:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:57 vm08.local ceph-mon[50506]: pgmap v2795: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:53:57.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:57 vm01.local ceph-mon[48597]: pgmap v2795: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:53:59.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:53:59 vm01.local ceph-mon[48597]: pgmap v2796: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:53:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:53:59 vm08.local ceph-mon[50506]: pgmap v2796: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:00.993 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:54:00.994 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:54:01.135 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:54:01.135 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:54:01.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:01 vm01.local ceph-mon[48597]: pgmap v2797: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:54:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:01 vm08.local ceph-mon[50506]: pgmap v2797: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:54:03.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:03 vm01.local ceph-mon[48597]: pgmap v2798: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:03.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:03 vm08.local ceph-mon[50506]: pgmap v2798: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:05 vm08.local ceph-mon[50506]: pgmap v2799: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:05 vm01.local ceph-mon[48597]: pgmap v2799: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:06.137 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:54:06.137 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:54:06.242 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:54:06.243 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:54:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:07 vm08.local ceph-mon[50506]: pgmap v2800: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:54:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:07 vm01.local ceph-mon[48597]: pgmap v2800: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:54:09.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:09 vm01.local ceph-mon[48597]: pgmap v2801: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:09 vm08.local ceph-mon[50506]: pgmap v2801: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:54:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:54:11.246 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:54:11.247 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:54:11.332 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:54:11.333 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:54:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:54:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:11 vm08.local ceph-mon[50506]: pgmap v2802: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:54:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:54:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:11 vm01.local ceph-mon[48597]: pgmap v2802: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:54:13.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:13 vm08.local ceph-mon[50506]: pgmap v2803: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:13.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:13 vm01.local ceph-mon[48597]: pgmap v2803: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:15.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:15 vm08.local ceph-mon[50506]: pgmap v2804: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:15.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:15 vm01.local ceph-mon[48597]: pgmap v2804: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:16.337 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:54:16.337 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:54:16.411 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:54:16.411 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:54:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:17 vm08.local ceph-mon[50506]: pgmap v2805: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:54:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:17 vm01.local ceph-mon[48597]: pgmap v2805: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:54:19.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:19 vm01.local ceph-mon[48597]: pgmap v2806: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:19.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:19 vm08.local ceph-mon[50506]: pgmap v2806: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:21.412 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:54:21.413 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:54:21.505 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:54:21.506 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:54:21.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:21 vm08.local ceph-mon[50506]: pgmap v2807: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:54:21.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:21 vm01.local ceph-mon[48597]: pgmap v2807: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:54:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:23 vm08.local ceph-mon[50506]: pgmap v2808: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:23.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:23 vm01.local ceph-mon[48597]: pgmap v2808: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:25.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:25 vm08.local ceph-mon[50506]: pgmap v2809: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:25 vm01.local ceph-mon[48597]: pgmap v2809: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:26.507 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:54:26.508 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:54:26.612 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:54:26.613 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:54:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:27 vm08.local ceph-mon[50506]: pgmap v2810: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:54:27.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:27 vm01.local ceph-mon[48597]: pgmap v2810: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:54:29.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:29 vm08.local ceph-mon[50506]: pgmap v2811: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:29.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:29 vm01.local ceph-mon[48597]: pgmap v2811: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:31.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:31 vm08.local ceph-mon[50506]: pgmap v2812: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:54:31.615 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:54:31.615 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:54:31.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:31 vm01.local ceph-mon[48597]: pgmap v2812: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:54:31.641 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:54:31.641 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:54:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:33 vm08.local ceph-mon[50506]: pgmap v2813: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:33.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:33 vm01.local ceph-mon[48597]: pgmap v2813: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:35.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:35 vm08.local ceph-mon[50506]: pgmap v2814: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:35.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:35 vm01.local ceph-mon[48597]: pgmap v2814: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:36.643 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:54:36.643 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:54:36.670 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:54:36.670 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:54:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:37 vm08.local ceph-mon[50506]: pgmap v2815: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:54:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:37 vm01.local ceph-mon[48597]: pgmap v2815: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:54:39.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:39 vm01.local ceph-mon[48597]: pgmap v2816: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:39 vm08.local ceph-mon[50506]: pgmap v2816: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:41.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:41 vm08.local ceph-mon[50506]: pgmap v2817: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:54:41.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:41 vm01.local ceph-mon[48597]: pgmap v2817: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:54:41.672 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:54:41.672 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:54:41.699 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:54:41.699 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:54:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:43 vm08.local ceph-mon[50506]: pgmap v2818: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:43.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:43 vm01.local ceph-mon[48597]: pgmap v2818: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:45.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:45 vm08.local ceph-mon[50506]: pgmap v2819: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:45.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:45 vm01.local ceph-mon[48597]: pgmap v2819: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:46.700 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:54:46.701 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:54:46.726 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:54:46.727 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:54:47.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:47 vm08.local ceph-mon[50506]: pgmap v2820: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:54:47.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:47 vm01.local ceph-mon[48597]: pgmap v2820: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:54:48.277 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:48 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:54:48.277 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:48 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:54:48.277 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:48 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:54:48.295 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:48 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:54:48.295 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:48 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:54:48.295 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:48 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:54:49.724 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:49 vm08.local ceph-mon[50506]: pgmap v2821: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:49.859 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:49 vm01.local ceph-mon[48597]: pgmap v2821: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:51.030 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:51 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:54:51.288 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:50 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:54:51.728 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:54:51.729 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:54:51.881 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:54:51.882 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:54:52.455 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:52 vm08.local ceph-mon[50506]: pgmap v2822: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:54:52.455 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:52 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:54:52.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:52 vm01.local ceph-mon[48597]: pgmap v2822: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:54:52.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:52 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:54:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:53 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:54:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:53 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:54:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:53 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:54:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:53 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:54:54.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:53 vm01.local ceph-mon[48597]: pgmap v2823: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:54.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:53 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:54:54.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:53 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:54:54.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:53 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:54:54.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:53 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:54:54.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:53 vm08.local ceph-mon[50506]: pgmap v2823: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:55.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:55 vm01.local ceph-mon[48597]: pgmap v2824: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:55 vm08.local ceph-mon[50506]: pgmap v2824: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:56.885 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:54:56.886 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:54:57.071 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:54:57.072 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:54:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:57 vm01.local ceph-mon[48597]: pgmap v2825: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:54:57.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:57 vm08.local ceph-mon[50506]: pgmap v2825: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:54:59.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:54:59 vm01.local ceph-mon[48597]: pgmap v2826: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:54:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:54:59 vm08.local ceph-mon[50506]: pgmap v2826: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:01.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:01 vm01.local ceph-mon[48597]: pgmap v2827: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:55:01.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:01 vm08.local ceph-mon[50506]: pgmap v2827: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:55:02.074 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:55:02.075 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:55:02.230 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:55:02.231 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:55:03.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:03 vm08.local ceph-mon[50506]: pgmap v2828: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:03.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:03 vm01.local ceph-mon[48597]: pgmap v2828: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:05 vm01.local ceph-mon[48597]: pgmap v2829: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:06.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:05 vm08.local ceph-mon[50506]: pgmap v2829: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:07.232 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:55:07.233 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:55:07.365 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:55:07.366 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:55:07.846 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:07 vm08.local ceph-mon[50506]: pgmap v2830: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:55:07.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:07 vm01.local ceph-mon[48597]: pgmap v2830: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:55:09.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:09 vm08.local ceph-mon[50506]: pgmap v2831: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:09.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:09 vm01.local ceph-mon[48597]: pgmap v2831: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:55:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:55:12.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:55:12.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:11 vm08.local ceph-mon[50506]: pgmap v2832: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:55:12.119 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:55:12.119 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:11 vm01.local ceph-mon[48597]: pgmap v2832: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:55:12.368 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:55:12.368 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:55:12.441 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:55:12.441 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:55:13.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:13 vm01.local ceph-mon[48597]: pgmap v2833: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:13 vm08.local ceph-mon[50506]: pgmap v2833: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:15.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:15 vm08.local ceph-mon[50506]: pgmap v2834: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:15.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:15 vm01.local ceph-mon[48597]: pgmap v2834: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:17.443 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:55:17.443 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:55:17.548 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:55:17.548 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:55:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:17 vm01.local ceph-mon[48597]: pgmap v2835: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:55:17.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:17 vm08.local ceph-mon[50506]: pgmap v2835: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:55:19.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:19 vm01.local ceph-mon[48597]: pgmap v2836: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:19.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:19 vm08.local ceph-mon[50506]: pgmap v2836: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:21.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:21 vm08.local ceph-mon[50506]: pgmap v2837: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:55:21.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:21 vm01.local ceph-mon[48597]: pgmap v2837: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:55:22.550 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:55:22.550 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:55:22.576 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:55:22.576 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:55:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:23 vm08.local ceph-mon[50506]: pgmap v2838: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:23.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:23 vm01.local ceph-mon[48597]: pgmap v2838: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:25.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:25 vm08.local ceph-mon[50506]: pgmap v2839: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:25 vm01.local ceph-mon[48597]: pgmap v2839: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:27.458 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:27 vm08.local ceph-mon[50506]: pgmap v2840: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:55:27.577 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:55:27.578 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:55:27.606 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:55:27.606 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:55:27.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:27 vm01.local ceph-mon[48597]: pgmap v2840: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:55:29.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:29 vm01.local ceph-mon[48597]: pgmap v2841: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:29 vm08.local ceph-mon[50506]: pgmap v2841: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:31.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:31 vm08.local ceph-mon[50506]: pgmap v2842: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:55:31.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:31 vm01.local ceph-mon[48597]: pgmap v2842: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:55:32.607 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:55:32.608 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:55:32.634 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:55:32.634 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:55:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:33 vm08.local ceph-mon[50506]: pgmap v2843: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:33.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:33 vm01.local ceph-mon[48597]: pgmap v2843: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:35.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:35 vm08.local ceph-mon[50506]: pgmap v2844: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:35.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:35 vm01.local ceph-mon[48597]: pgmap v2844: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:37 vm08.local ceph-mon[50506]: pgmap v2845: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:55:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:37 vm01.local ceph-mon[48597]: pgmap v2845: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:55:37.635 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:55:37.636 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:55:37.662 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:55:37.662 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:55:39.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:39 vm01.local ceph-mon[48597]: pgmap v2846: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:39 vm08.local ceph-mon[50506]: pgmap v2846: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:41.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:41 vm08.local ceph-mon[50506]: pgmap v2847: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:55:41.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:41 vm01.local ceph-mon[48597]: pgmap v2847: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:55:42.664 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:55:42.664 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:55:42.691 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:55:42.691 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:55:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:43 vm08.local ceph-mon[50506]: pgmap v2848: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:43.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:43 vm01.local ceph-mon[48597]: pgmap v2848: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:45.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:45 vm08.local ceph-mon[50506]: pgmap v2849: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:45.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:45 vm01.local ceph-mon[48597]: pgmap v2849: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:47.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:47 vm08.local ceph-mon[50506]: pgmap v2850: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:55:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:47 vm01.local ceph-mon[48597]: pgmap v2850: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:55:47.692 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:55:47.693 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:55:47.718 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:55:47.719 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:55:49.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:49 vm01.local ceph-mon[48597]: pgmap v2851: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:49.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:49 vm08.local ceph-mon[50506]: pgmap v2851: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:51.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:51 vm08.local ceph-mon[50506]: pgmap v2852: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:55:51.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:51 vm01.local ceph-mon[48597]: pgmap v2852: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:55:52.720 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:55:52.721 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:55:52.746 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:55:52.746 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:55:53.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:52 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:55:53.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:52 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:55:53.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:52 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:55:53.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:52 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:55:53.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:52 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:55:53.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:52 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:55:54.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:54 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:55:54.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:54 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:55:54.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:54 vm01.local ceph-mon[48597]: pgmap v2853: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:54.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:54 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:55:54.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:54 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:55:54.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:54 vm08.local ceph-mon[50506]: pgmap v2853: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:55.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:55 vm08.local ceph-mon[50506]: pgmap v2854: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:55.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:55 vm01.local ceph-mon[48597]: pgmap v2854: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:57 vm08.local ceph-mon[50506]: pgmap v2855: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:55:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:57 vm01.local ceph-mon[48597]: pgmap v2855: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:55:57.747 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:55:57.748 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:55:57.774 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:55:57.775 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:55:59.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:55:59 vm01.local ceph-mon[48597]: pgmap v2856: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:55:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:55:59 vm08.local ceph-mon[50506]: pgmap v2856: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:01.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:01 vm08.local ceph-mon[50506]: pgmap v2857: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:56:01.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:01 vm01.local ceph-mon[48597]: pgmap v2857: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:56:02.776 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:56:02.776 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:56:02.813 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:56:02.813 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:56:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:03 vm08.local ceph-mon[50506]: pgmap v2858: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:03.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:03 vm01.local ceph-mon[48597]: pgmap v2858: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:05 vm08.local ceph-mon[50506]: pgmap v2859: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:05 vm01.local ceph-mon[48597]: pgmap v2859: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:07 vm08.local ceph-mon[50506]: pgmap v2860: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:56:07.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:07 vm01.local ceph-mon[48597]: pgmap v2860: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:56:07.814 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:56:07.815 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:56:07.841 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:56:07.841 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:56:09.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:09 vm01.local ceph-mon[48597]: pgmap v2861: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:09 vm08.local ceph-mon[50506]: pgmap v2861: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:10.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:56:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:56:11.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:56:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:11 vm08.local ceph-mon[50506]: pgmap v2862: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:56:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:56:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:11 vm01.local ceph-mon[48597]: pgmap v2862: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:56:12.842 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:56:12.843 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:56:12.870 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:56:12.870 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:56:13.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:13 vm08.local ceph-mon[50506]: pgmap v2863: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:13.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:13 vm01.local ceph-mon[48597]: pgmap v2863: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:15.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:15 vm08.local ceph-mon[50506]: pgmap v2864: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:15.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:15 vm01.local ceph-mon[48597]: pgmap v2864: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:17.593 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:17 vm08.local ceph-mon[50506]: pgmap v2865: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:56:17.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:17 vm01.local ceph-mon[48597]: pgmap v2865: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:56:17.872 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:56:17.872 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:56:17.898 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:56:17.899 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:56:19.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:19 vm01.local ceph-mon[48597]: pgmap v2866: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:19.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:19 vm08.local ceph-mon[50506]: pgmap v2866: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:21.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:21 vm08.local ceph-mon[50506]: pgmap v2867: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:56:21.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:21 vm01.local ceph-mon[48597]: pgmap v2867: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:56:22.900 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:56:22.901 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:56:22.928 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:56:22.929 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:56:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:23 vm08.local ceph-mon[50506]: pgmap v2868: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:23.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:23 vm01.local ceph-mon[48597]: pgmap v2868: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:25.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:25 vm08.local ceph-mon[50506]: pgmap v2869: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:25 vm01.local ceph-mon[48597]: pgmap v2869: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:27 vm08.local ceph-mon[50506]: pgmap v2870: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:56:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:27 vm01.local ceph-mon[48597]: pgmap v2870: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:56:27.930 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:56:27.931 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:56:27.956 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:56:27.956 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:56:29.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:29 vm01.local ceph-mon[48597]: pgmap v2871: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:29 vm08.local ceph-mon[50506]: pgmap v2871: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:31.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:31 vm08.local ceph-mon[50506]: pgmap v2872: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:56:31.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:31 vm01.local ceph-mon[48597]: pgmap v2872: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:56:32.958 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:56:32.958 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:56:32.985 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:56:32.986 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:56:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:33 vm08.local ceph-mon[50506]: pgmap v2873: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:33.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:33 vm01.local ceph-mon[48597]: pgmap v2873: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:35.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:35 vm08.local ceph-mon[50506]: pgmap v2874: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:35.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:35 vm01.local ceph-mon[48597]: pgmap v2874: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:37 vm08.local ceph-mon[50506]: pgmap v2875: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:56:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:37 vm01.local ceph-mon[48597]: pgmap v2875: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:56:37.987 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:56:37.987 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:56:38.014 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:56:38.014 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:56:39.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:39 vm01.local ceph-mon[48597]: pgmap v2876: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:39 vm08.local ceph-mon[50506]: pgmap v2876: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:41.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:41 vm08.local ceph-mon[50506]: pgmap v2877: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:56:41.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:41 vm01.local ceph-mon[48597]: pgmap v2877: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:56:43.015 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:56:43.016 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:56:43.042 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:56:43.042 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:56:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:43 vm08.local ceph-mon[50506]: pgmap v2878: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:43.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:43 vm01.local ceph-mon[48597]: pgmap v2878: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:45.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:45 vm08.local ceph-mon[50506]: pgmap v2879: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:45.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:45 vm01.local ceph-mon[48597]: pgmap v2879: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:47.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:47 vm08.local ceph-mon[50506]: pgmap v2880: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:56:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:47 vm01.local ceph-mon[48597]: pgmap v2880: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:56:48.043 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:56:48.044 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:56:48.070 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:56:48.071 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:56:49.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:49 vm01.local ceph-mon[48597]: pgmap v2881: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:49.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:49 vm08.local ceph-mon[50506]: pgmap v2881: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:51.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:51 vm08.local ceph-mon[50506]: pgmap v2882: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:56:51.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:51 vm01.local ceph-mon[48597]: pgmap v2882: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:56:53.072 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:56:53.072 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:56:53.097 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:56:53.098 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:56:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:53 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:56:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:53 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:56:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:53 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:56:53.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:53 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:56:53.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:53 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:56:53.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:53 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:56:54.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:54 vm01.local ceph-mon[48597]: pgmap v2883: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:54.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:54 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:56:54.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:54 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:56:54.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:54 vm08.local ceph-mon[50506]: pgmap v2883: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:54.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:54 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:56:54.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:54 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:56:55.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:55 vm08.local ceph-mon[50506]: pgmap v2884: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:55.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:55 vm01.local ceph-mon[48597]: pgmap v2884: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:57 vm08.local ceph-mon[50506]: pgmap v2885: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:56:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:57 vm01.local ceph-mon[48597]: pgmap v2885: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:56:58.099 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:56:58.100 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:56:58.126 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:56:58.127 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:56:59.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:56:59 vm01.local ceph-mon[48597]: pgmap v2886: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:56:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:56:59 vm08.local ceph-mon[50506]: pgmap v2886: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:01.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:01 vm08.local ceph-mon[50506]: pgmap v2887: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:57:01.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:01 vm01.local ceph-mon[48597]: pgmap v2887: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:57:03.128 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:57:03.129 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:57:03.155 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:57:03.155 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:57:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:03 vm08.local ceph-mon[50506]: pgmap v2888: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:03.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:03 vm01.local ceph-mon[48597]: pgmap v2888: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:05 vm08.local ceph-mon[50506]: pgmap v2889: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:05.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:05 vm01.local ceph-mon[48597]: pgmap v2889: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:07 vm08.local ceph-mon[50506]: pgmap v2890: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:57:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:07 vm01.local ceph-mon[48597]: pgmap v2890: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:57:08.157 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:57:08.157 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:57:08.183 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:57:08.183 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:57:09.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:09 vm01.local ceph-mon[48597]: pgmap v2891: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:09 vm08.local ceph-mon[50506]: pgmap v2891: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:57:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:57:11.843 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:57:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:11 vm08.local ceph-mon[50506]: pgmap v2892: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:57:11.867 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:57:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:11 vm01.local ceph-mon[48597]: pgmap v2892: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:57:13.185 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:57:13.185 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:57:13.211 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:57:13.212 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:57:13.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:13 vm08.local ceph-mon[50506]: pgmap v2893: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:13.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:13 vm01.local ceph-mon[48597]: pgmap v2893: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:15.845 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:15 vm08.local ceph-mon[50506]: pgmap v2894: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:15.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:15 vm01.local ceph-mon[48597]: pgmap v2894: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:17 vm08.local ceph-mon[50506]: pgmap v2895: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:57:17.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:17 vm01.local ceph-mon[48597]: pgmap v2895: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:57:18.213 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:57:18.214 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:57:18.240 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:57:18.241 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:57:19.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:19 vm01.local ceph-mon[48597]: pgmap v2896: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:19.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:19 vm08.local ceph-mon[50506]: pgmap v2896: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:21.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:21 vm08.local ceph-mon[50506]: pgmap v2897: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:57:21.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:21 vm01.local ceph-mon[48597]: pgmap v2897: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:57:23.242 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:57:23.243 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:57:23.270 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:57:23.270 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:57:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:23 vm08.local ceph-mon[50506]: pgmap v2898: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:23.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:23 vm01.local ceph-mon[48597]: pgmap v2898: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:25.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:25 vm08.local ceph-mon[50506]: pgmap v2899: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:25 vm01.local ceph-mon[48597]: pgmap v2899: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:27 vm08.local ceph-mon[50506]: pgmap v2900: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:57:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:27 vm01.local ceph-mon[48597]: pgmap v2900: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:57:28.272 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:57:28.273 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:57:28.302 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:57:28.302 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:57:29.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:29 vm01.local ceph-mon[48597]: pgmap v2901: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:29 vm08.local ceph-mon[50506]: pgmap v2901: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:31.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:31 vm08.local ceph-mon[50506]: pgmap v2902: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:57:31.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:31 vm01.local ceph-mon[48597]: pgmap v2902: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:57:33.304 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:57:33.304 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:57:33.330 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:57:33.331 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:57:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:33 vm08.local ceph-mon[50506]: pgmap v2903: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:33.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:33 vm01.local ceph-mon[48597]: pgmap v2903: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:35.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:35 vm08.local ceph-mon[50506]: pgmap v2904: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:35.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:35 vm01.local ceph-mon[48597]: pgmap v2904: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:37 vm08.local ceph-mon[50506]: pgmap v2905: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:57:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:37 vm01.local ceph-mon[48597]: pgmap v2905: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:57:38.332 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:57:38.333 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:57:38.359 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:57:38.360 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:57:39.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:39 vm01.local ceph-mon[48597]: pgmap v2906: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:39 vm08.local ceph-mon[50506]: pgmap v2906: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:41.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:41 vm08.local ceph-mon[50506]: pgmap v2907: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:57:41.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:41 vm01.local ceph-mon[48597]: pgmap v2907: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:57:43.361 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:57:43.361 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:57:43.392 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:57:43.393 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:57:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:43 vm08.local ceph-mon[50506]: pgmap v2908: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:43.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:43 vm01.local ceph-mon[48597]: pgmap v2908: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:45.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:45 vm08.local ceph-mon[50506]: pgmap v2909: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:45.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:45 vm01.local ceph-mon[48597]: pgmap v2909: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:47.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:47 vm08.local ceph-mon[50506]: pgmap v2910: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:57:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:47 vm01.local ceph-mon[48597]: pgmap v2910: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:57:48.394 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:57:48.395 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:57:48.430 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:57:48.431 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:57:49.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:49 vm01.local ceph-mon[48597]: pgmap v2911: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:49.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:49 vm08.local ceph-mon[50506]: pgmap v2911: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:51.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:51 vm08.local ceph-mon[50506]: pgmap v2912: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:57:51.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:51 vm01.local ceph-mon[48597]: pgmap v2912: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:57:53.432 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:57:53.433 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:57:53.468 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:57:53.468 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:57:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:53 vm08.local ceph-mon[50506]: pgmap v2913: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:53.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:53 vm01.local ceph-mon[48597]: pgmap v2913: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:54.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:54 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:57:54.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:54 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:57:54.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:54 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:57:54.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:54 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:57:54.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:54 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:57:54.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:54 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:57:54.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:54 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:57:54.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:54 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:57:54.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:54 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:57:54.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:54 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:57:55.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:55 vm08.local ceph-mon[50506]: pgmap v2914: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:55.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:55 vm01.local ceph-mon[48597]: pgmap v2914: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:57 vm08.local ceph-mon[50506]: pgmap v2915: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:57:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:57 vm01.local ceph-mon[48597]: pgmap v2915: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:57:58.472 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:57:58.472 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:57:58.499 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:57:58.500 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:57:59.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:57:59 vm01.local ceph-mon[48597]: pgmap v2916: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:57:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:57:59 vm08.local ceph-mon[50506]: pgmap v2916: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:01.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:01 vm08.local ceph-mon[50506]: pgmap v2917: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:58:01.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:01 vm01.local ceph-mon[48597]: pgmap v2917: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:58:03.502 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:58:03.502 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:58:03.528 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:58:03.528 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:58:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:03 vm08.local ceph-mon[50506]: pgmap v2918: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:03.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:03 vm01.local ceph-mon[48597]: pgmap v2918: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:05.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:05 vm08.local ceph-mon[50506]: pgmap v2919: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:05.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:05 vm01.local ceph-mon[48597]: pgmap v2919: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:07 vm08.local ceph-mon[50506]: pgmap v2920: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:58:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:07 vm01.local ceph-mon[48597]: pgmap v2920: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:58:08.530 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:58:08.530 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:58:08.557 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:58:08.558 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:58:09.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:09 vm01.local ceph-mon[48597]: pgmap v2921: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:09 vm08.local ceph-mon[50506]: pgmap v2921: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:58:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:58:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:58:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:58:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:11 vm08.local ceph-mon[50506]: pgmap v2922: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:58:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:11 vm01.local ceph-mon[48597]: pgmap v2922: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:58:13.559 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:58:13.559 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:58:13.586 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:58:13.586 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:58:13.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:13 vm08.local ceph-mon[50506]: pgmap v2923: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:13.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:13 vm01.local ceph-mon[48597]: pgmap v2923: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:16.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:15 vm08.local ceph-mon[50506]: pgmap v2924: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:16.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:15 vm01.local ceph-mon[48597]: pgmap v2924: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:17 vm08.local ceph-mon[50506]: pgmap v2925: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:58:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:17 vm01.local ceph-mon[48597]: pgmap v2925: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:58:18.588 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:58:18.588 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:58:18.615 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:58:18.615 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:58:19.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:19 vm01.local ceph-mon[48597]: pgmap v2926: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:19.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:19 vm08.local ceph-mon[50506]: pgmap v2926: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:21.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:21 vm08.local ceph-mon[50506]: pgmap v2927: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:58:21.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:21 vm01.local ceph-mon[48597]: pgmap v2927: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:58:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:23 vm08.local ceph-mon[50506]: pgmap v2928: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:23.617 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:58:23.617 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:58:23.617 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:23 vm01.local ceph-mon[48597]: pgmap v2928: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:23.644 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:58:23.644 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:58:25.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:25 vm08.local ceph-mon[50506]: pgmap v2929: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:25 vm01.local ceph-mon[48597]: pgmap v2929: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:27 vm08.local ceph-mon[50506]: pgmap v2930: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:58:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:27 vm01.local ceph-mon[48597]: pgmap v2930: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:58:28.646 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:58:28.646 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:58:28.673 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:58:28.673 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:58:29.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:29 vm01.local ceph-mon[48597]: pgmap v2931: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:29 vm08.local ceph-mon[50506]: pgmap v2931: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:31.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:31 vm08.local ceph-mon[50506]: pgmap v2932: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:58:31.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:31 vm01.local ceph-mon[48597]: pgmap v2932: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:58:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:33 vm08.local ceph-mon[50506]: pgmap v2933: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:33.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:33 vm01.local ceph-mon[48597]: pgmap v2933: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:33.675 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:58:33.675 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:58:33.702 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:58:33.703 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:58:35.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:35 vm08.local ceph-mon[50506]: pgmap v2934: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:35.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:35 vm01.local ceph-mon[48597]: pgmap v2934: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:37 vm08.local ceph-mon[50506]: pgmap v2935: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:58:37.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:37 vm01.local ceph-mon[48597]: pgmap v2935: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:58:38.704 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:58:38.705 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:58:38.731 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:58:38.732 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:58:39.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:39 vm01.local ceph-mon[48597]: pgmap v2936: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:39 vm08.local ceph-mon[50506]: pgmap v2936: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:41.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:41 vm08.local ceph-mon[50506]: pgmap v2937: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:58:41.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:41 vm01.local ceph-mon[48597]: pgmap v2937: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:58:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:43 vm08.local ceph-mon[50506]: pgmap v2938: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:43.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:43 vm01.local ceph-mon[48597]: pgmap v2938: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:43.733 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:58:43.734 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:58:43.762 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:58:43.762 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:58:46.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:46 vm08.local ceph-mon[50506]: pgmap v2939: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:46.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:46 vm01.local ceph-mon[48597]: pgmap v2939: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:48.764 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:58:48.764 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:58:48.792 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:58:48.792 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:58:48.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:48 vm08.local ceph-mon[50506]: pgmap v2940: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:58:48.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:48 vm01.local ceph-mon[48597]: pgmap v2940: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:58:50.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:50 vm08.local ceph-mon[50506]: pgmap v2941: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:50.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:50 vm01.local ceph-mon[48597]: pgmap v2941: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:52.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:52 vm08.local ceph-mon[50506]: pgmap v2942: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:58:52.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:52 vm01.local ceph-mon[48597]: pgmap v2942: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:58:53.794 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:58:53.794 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:58:53.830 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:58:53.830 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:58:54.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:54 vm08.local ceph-mon[50506]: pgmap v2943: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:54.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:54 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:58:54.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:54 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:58:54.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:54 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:58:54.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:54 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:58:54.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:54 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:58:54.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:54 vm01.local ceph-mon[48597]: pgmap v2943: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:54.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:54 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:58:54.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:54 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:58:54.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:54 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:58:54.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:54 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:58:54.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:54 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:58:56.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:56 vm08.local ceph-mon[50506]: pgmap v2944: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:56.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:56 vm01.local ceph-mon[48597]: pgmap v2944: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:58:58.832 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:58:58.833 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:58:58.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:58:58 vm08.local ceph-mon[50506]: pgmap v2945: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:58:58.860 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:58:58.860 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:58:58.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:58:58 vm01.local ceph-mon[48597]: pgmap v2945: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:59:00.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:00 vm08.local ceph-mon[50506]: pgmap v2946: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:00.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:00 vm01.local ceph-mon[48597]: pgmap v2946: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:02.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:02 vm08.local ceph-mon[50506]: pgmap v2947: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:59:02.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:02 vm01.local ceph-mon[48597]: pgmap v2947: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:59:03.862 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:59:03.862 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:59:03.888 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:59:03.889 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:59:04.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:04 vm08.local ceph-mon[50506]: pgmap v2948: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:04.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:04 vm01.local ceph-mon[48597]: pgmap v2948: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:06.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:06 vm08.local ceph-mon[50506]: pgmap v2949: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:06.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:06 vm01.local ceph-mon[48597]: pgmap v2949: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:08.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:08 vm08.local ceph-mon[50506]: pgmap v2950: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:59:08.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:08 vm01.local ceph-mon[48597]: pgmap v2950: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:59:08.890 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:59:08.890 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:59:08.916 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:59:08.917 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:59:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:10 vm08.local ceph-mon[50506]: pgmap v2951: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:59:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:59:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:10 vm01.local ceph-mon[48597]: pgmap v2951: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:59:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T11:59:12.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:12 vm08.local ceph-mon[50506]: pgmap v2952: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:59:12.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:12 vm01.local ceph-mon[48597]: pgmap v2952: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:59:13.918 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:59:13.919 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:59:13.948 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:59:13.948 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:59:14.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:14 vm08.local ceph-mon[50506]: pgmap v2953: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:14 vm01.local ceph-mon[48597]: pgmap v2953: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:16.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:16 vm08.local ceph-mon[50506]: pgmap v2954: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:16.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:16 vm01.local ceph-mon[48597]: pgmap v2954: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:18.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:18 vm01.local ceph-mon[48597]: pgmap v2955: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:59:18.950 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:59:18.950 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:59:18.977 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:59:18.978 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:59:19.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:18 vm08.local ceph-mon[50506]: pgmap v2955: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:59:20.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:20 vm01.local ceph-mon[48597]: pgmap v2956: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:21.093 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:20 vm08.local ceph-mon[50506]: pgmap v2956: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:22.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:22 vm01.local ceph-mon[48597]: pgmap v2957: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:59:23.093 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:22 vm08.local ceph-mon[50506]: pgmap v2957: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:59:23.979 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:59:23.980 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:59:24.006 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:59:24.006 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:59:24.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:24 vm01.local ceph-mon[48597]: pgmap v2958: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:25.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:24 vm08.local ceph-mon[50506]: pgmap v2958: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:26.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:26 vm08.local ceph-mon[50506]: pgmap v2959: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:27.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:26 vm01.local ceph-mon[48597]: pgmap v2959: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:29.007 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:59:29.008 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:59:29.035 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:59:29.036 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:59:29.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:28 vm08.local ceph-mon[50506]: pgmap v2960: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:59:29.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:28 vm01.local ceph-mon[48597]: pgmap v2960: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:59:30.992 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:30 vm01.local ceph-mon[48597]: pgmap v2961: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:31.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:30 vm08.local ceph-mon[50506]: pgmap v2961: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:33.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:32 vm08.local ceph-mon[50506]: pgmap v2962: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:59:33.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:32 vm01.local ceph-mon[48597]: pgmap v2962: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:59:34.037 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:59:34.038 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:59:34.065 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:59:34.065 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:59:35.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:34 vm08.local ceph-mon[50506]: pgmap v2963: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:35.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:34 vm01.local ceph-mon[48597]: pgmap v2963: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:36.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:36 vm08.local ceph-mon[50506]: pgmap v2964: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:37.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:36 vm01.local ceph-mon[48597]: pgmap v2964: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:39.067 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:59:39.067 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:59:39.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:38 vm08.local ceph-mon[50506]: pgmap v2965: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:59:39.094 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:59:39.095 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:59:39.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:38 vm01.local ceph-mon[48597]: pgmap v2965: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:59:40.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:40 vm01.local ceph-mon[48597]: pgmap v2966: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:41.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:40 vm08.local ceph-mon[50506]: pgmap v2966: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:43.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:42 vm08.local ceph-mon[50506]: pgmap v2967: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:59:43.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:42 vm01.local ceph-mon[48597]: pgmap v2967: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:59:44.096 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:59:44.096 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:59:44.124 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:59:44.124 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:59:45.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:44 vm08.local ceph-mon[50506]: pgmap v2968: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:45.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:44 vm01.local ceph-mon[48597]: pgmap v2968: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:46.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:46 vm08.local ceph-mon[50506]: pgmap v2969: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:47.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:46 vm01.local ceph-mon[48597]: pgmap v2969: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:49.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:48 vm08.local ceph-mon[50506]: pgmap v2970: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:59:49.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:48 vm01.local ceph-mon[48597]: pgmap v2970: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:59:49.125 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:59:49.126 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:59:49.156 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:59:49.157 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:59:50.992 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:50 vm01.local ceph-mon[48597]: pgmap v2971: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:51.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:50 vm08.local ceph-mon[50506]: pgmap v2971: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:53.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:52 vm08.local ceph-mon[50506]: pgmap v2972: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:59:53.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:52 vm01.local ceph-mon[48597]: pgmap v2972: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:59:54.158 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:59:54.158 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:59:54.185 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:59:54.185 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T11:59:55.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:54 vm08.local ceph-mon[50506]: pgmap v2973: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:55.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:54 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:59:55.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:54 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:59:55.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:54 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:59:55.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:54 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:59:55.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:54 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:59:55.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:54 vm01.local ceph-mon[48597]: pgmap v2973: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:55.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:54 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:59:55.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:54 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:59:55.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:54 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:59:55.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:54 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:59:55.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:54 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T11:59:56.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:56 vm08.local ceph-mon[50506]: pgmap v2974: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:57.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:56 vm01.local ceph-mon[48597]: pgmap v2974: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T11:59:59.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 11:59:58 vm08.local ceph-mon[50506]: pgmap v2975: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:59:59.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 11:59:58 vm01.local ceph-mon[48597]: pgmap v2975: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T11:59:59.186 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T11:59:59.187 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T11:59:59.214 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T11:59:59.214 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:00:00.992 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:00 vm01.local ceph-mon[48597]: pgmap v2976: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:00.992 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:00 vm01.local ceph-mon[48597]: overall HEALTH_OK 2026-03-10T12:00:01.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:00 vm08.local ceph-mon[50506]: pgmap v2976: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:01.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:00 vm08.local ceph-mon[50506]: overall HEALTH_OK 2026-03-10T12:00:03.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:02 vm08.local ceph-mon[50506]: pgmap v2977: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:00:03.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:02 vm01.local ceph-mon[48597]: pgmap v2977: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:00:04.216 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:00:04.216 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:00:04.242 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:00:04.242 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:00:05.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:04 vm08.local ceph-mon[50506]: pgmap v2978: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:05.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:04 vm01.local ceph-mon[48597]: pgmap v2978: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:07.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:06 vm08.local ceph-mon[50506]: pgmap v2979: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:07.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:06 vm01.local ceph-mon[48597]: pgmap v2979: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:09.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:08 vm08.local ceph-mon[50506]: pgmap v2980: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:00:09.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:08 vm01.local ceph-mon[48597]: pgmap v2980: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:00:09.243 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:00:09.244 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:00:09.269 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:00:09.270 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:00:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:10 vm08.local ceph-mon[50506]: pgmap v2981: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:00:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:00:11.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:10 vm01.local ceph-mon[48597]: pgmap v2981: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:11.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:00:11.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:00:13.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:12 vm08.local ceph-mon[50506]: pgmap v2982: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:00:13.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:12 vm01.local ceph-mon[48597]: pgmap v2982: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:00:14.271 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:00:14.271 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:00:14.297 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:00:14.297 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:00:15.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:15 vm08.local ceph-mon[50506]: pgmap v2983: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:15.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:15 vm01.local ceph-mon[48597]: pgmap v2983: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:17.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:17 vm08.local ceph-mon[50506]: pgmap v2984: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:17.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:17 vm01.local ceph-mon[48597]: pgmap v2984: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:19.298 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:00:19.299 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:00:19.324 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:00:19.325 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:00:19.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:19 vm08.local ceph-mon[50506]: pgmap v2985: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:00:19.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:19 vm01.local ceph-mon[48597]: pgmap v2985: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:00:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:21 vm08.local ceph-mon[50506]: pgmap v2986: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:21 vm01.local ceph-mon[48597]: pgmap v2986: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:23.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:23 vm01.local ceph-mon[48597]: pgmap v2987: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:00:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:23 vm08.local ceph-mon[50506]: pgmap v2987: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:00:24.326 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:00:24.327 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:00:24.353 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:00:24.354 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:00:25.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:25 vm01.local ceph-mon[48597]: pgmap v2988: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:25.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:25 vm08.local ceph-mon[50506]: pgmap v2988: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:27.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:27 vm01.local ceph-mon[48597]: pgmap v2989: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:27 vm08.local ceph-mon[50506]: pgmap v2989: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:29.355 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:00:29.356 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:00:29.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:29 vm01.local ceph-mon[48597]: pgmap v2990: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:00:29.383 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:00:29.383 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:00:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:29 vm08.local ceph-mon[50506]: pgmap v2990: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:00:31.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:31 vm01.local ceph-mon[48597]: pgmap v2991: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:31.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:31 vm08.local ceph-mon[50506]: pgmap v2991: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:33 vm08.local ceph-mon[50506]: pgmap v2992: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:00:33.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:33 vm01.local ceph-mon[48597]: pgmap v2992: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:00:34.385 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:00:34.385 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:00:34.411 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:00:34.412 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:00:35.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:35 vm08.local ceph-mon[50506]: pgmap v2993: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:35.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:35 vm01.local ceph-mon[48597]: pgmap v2993: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:37 vm08.local ceph-mon[50506]: pgmap v2994: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:37 vm01.local ceph-mon[48597]: pgmap v2994: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:38.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:38 vm08.local ceph-mon[50506]: pgmap v2995: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:00:38.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:38 vm01.local ceph-mon[48597]: pgmap v2995: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:00:39.413 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:00:39.414 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:00:39.440 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:00:39.441 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:00:40.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:40 vm01.local ceph-mon[48597]: pgmap v2996: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:41.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:40 vm08.local ceph-mon[50506]: pgmap v2996: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:43.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:42 vm08.local ceph-mon[50506]: pgmap v2997: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:00:43.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:42 vm01.local ceph-mon[48597]: pgmap v2997: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:00:44.442 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:00:44.443 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:00:44.471 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:00:44.472 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:00:45.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:44 vm08.local ceph-mon[50506]: pgmap v2998: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:45.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:44 vm01.local ceph-mon[48597]: pgmap v2998: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:46.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:46 vm08.local ceph-mon[50506]: pgmap v2999: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:47.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:46 vm01.local ceph-mon[48597]: pgmap v2999: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:49.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:48 vm08.local ceph-mon[50506]: pgmap v3000: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:00:49.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:48 vm01.local ceph-mon[48597]: pgmap v3000: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:00:49.473 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:00:49.474 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:00:49.502 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:00:49.503 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:00:51.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:50 vm08.local ceph-mon[50506]: pgmap v3001: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:51.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:50 vm01.local ceph-mon[48597]: pgmap v3001: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:53.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:52 vm08.local ceph-mon[50506]: pgmap v3002: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:00:53.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:52 vm01.local ceph-mon[48597]: pgmap v3002: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:00:54.504 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:00:54.504 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:00:54.571 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:00:54.571 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:00:55.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:54 vm08.local ceph-mon[50506]: pgmap v3003: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:55.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:54 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:00:55.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:54 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:00:55.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:54 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:00:55.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:54 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:00:55.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:54 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:00:55.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:54 vm01.local ceph-mon[48597]: pgmap v3003: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:55.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:54 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:00:55.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:54 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:00:55.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:54 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:00:55.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:54 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:00:55.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:54 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:00:57.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:56 vm08.local ceph-mon[50506]: pgmap v3004: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:57.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:56 vm01.local ceph-mon[48597]: pgmap v3004: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:00:59.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:00:58 vm08.local ceph-mon[50506]: pgmap v3005: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:00:59.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:00:58 vm01.local ceph-mon[48597]: pgmap v3005: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:00:59.573 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:00:59.573 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:00:59.600 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:00:59.600 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:01:01.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:00 vm08.local ceph-mon[50506]: pgmap v3006: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:01.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:00 vm01.local ceph-mon[48597]: pgmap v3006: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:03.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:02 vm08.local ceph-mon[50506]: pgmap v3007: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:01:03.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:02 vm01.local ceph-mon[48597]: pgmap v3007: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:01:04.602 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:01:04.602 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:01:04.629 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:01:04.630 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:01:05.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:04 vm08.local ceph-mon[50506]: pgmap v3008: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:05.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:04 vm01.local ceph-mon[48597]: pgmap v3008: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:07.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:06 vm08.local ceph-mon[50506]: pgmap v3009: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:07.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:06 vm01.local ceph-mon[48597]: pgmap v3009: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:09.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:08 vm08.local ceph-mon[50506]: pgmap v3010: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:01:09.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:08 vm01.local ceph-mon[48597]: pgmap v3010: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:01:09.631 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:01:09.631 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:01:09.657 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:01:09.657 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:01:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:10 vm08.local ceph-mon[50506]: pgmap v3011: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:01:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:01:11.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:10 vm01.local ceph-mon[48597]: pgmap v3011: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:11.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:01:11.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:01:13.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:12 vm08.local ceph-mon[50506]: pgmap v3012: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:01:13.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:12 vm01.local ceph-mon[48597]: pgmap v3012: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:01:14.658 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:01:14.659 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:01:14.687 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:01:14.687 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:01:15.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:14 vm08.local ceph-mon[50506]: pgmap v3013: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:15.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:14 vm01.local ceph-mon[48597]: pgmap v3013: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:17.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:16 vm08.local ceph-mon[50506]: pgmap v3014: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:17.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:16 vm01.local ceph-mon[48597]: pgmap v3014: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:19.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:18 vm08.local ceph-mon[50506]: pgmap v3015: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:01:19.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:18 vm01.local ceph-mon[48597]: pgmap v3015: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:01:19.689 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:01:19.690 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:01:19.717 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:01:19.718 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:01:21.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:20 vm08.local ceph-mon[50506]: pgmap v3016: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:21.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:20 vm01.local ceph-mon[48597]: pgmap v3016: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:23.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:22 vm08.local ceph-mon[50506]: pgmap v3017: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:01:23.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:22 vm01.local ceph-mon[48597]: pgmap v3017: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:01:24.719 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:01:24.720 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:01:24.746 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:01:24.747 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:01:25.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:24 vm08.local ceph-mon[50506]: pgmap v3018: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:25.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:24 vm01.local ceph-mon[48597]: pgmap v3018: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:27.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:26 vm01.local ceph-mon[48597]: pgmap v3019: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:27.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:26 vm08.local ceph-mon[50506]: pgmap v3019: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:29.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:28 vm01.local ceph-mon[48597]: pgmap v3020: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:01:29.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:28 vm08.local ceph-mon[50506]: pgmap v3020: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:01:29.748 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:01:29.748 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:01:29.775 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:01:29.776 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:01:31.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:30 vm01.local ceph-mon[48597]: pgmap v3021: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:31.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:30 vm08.local ceph-mon[50506]: pgmap v3021: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:33.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:32 vm01.local ceph-mon[48597]: pgmap v3022: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:01:33.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:32 vm08.local ceph-mon[50506]: pgmap v3022: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:01:34.778 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:01:34.779 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:01:34.809 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:01:34.810 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:01:35.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:34 vm08.local ceph-mon[50506]: pgmap v3023: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:35.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:34 vm01.local ceph-mon[48597]: pgmap v3023: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:37.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:36 vm08.local ceph-mon[50506]: pgmap v3024: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:37.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:36 vm01.local ceph-mon[48597]: pgmap v3024: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:39.132 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:38 vm01.local ceph-mon[48597]: pgmap v3025: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:01:39.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:38 vm08.local ceph-mon[50506]: pgmap v3025: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:01:39.812 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:01:39.812 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:01:39.964 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:01:39.965 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:01:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:41 vm01.local ceph-mon[48597]: pgmap v3026: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:41.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:41 vm08.local ceph-mon[50506]: pgmap v3026: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:43 vm08.local ceph-mon[50506]: pgmap v3027: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:01:43.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:43 vm01.local ceph-mon[48597]: pgmap v3027: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:01:44.966 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:01:44.967 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:01:44.994 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:01:44.995 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:01:45.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:45 vm08.local ceph-mon[50506]: pgmap v3028: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:45.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:45 vm01.local ceph-mon[48597]: pgmap v3028: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:47.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:47 vm08.local ceph-mon[50506]: pgmap v3029: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:47 vm01.local ceph-mon[48597]: pgmap v3029: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:49.420 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:49 vm01.local ceph-mon[48597]: pgmap v3030: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:01:49.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:49 vm08.local ceph-mon[50506]: pgmap v3030: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:01:49.996 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:01:49.996 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:01:50.022 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:01:50.023 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:01:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:51 vm01.local ceph-mon[48597]: pgmap v3031: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:51.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:51 vm08.local ceph-mon[50506]: pgmap v3031: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:53 vm08.local ceph-mon[50506]: pgmap v3032: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:01:53.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:53 vm01.local ceph-mon[48597]: pgmap v3032: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:01:55.024 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:01:55.025 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:01:55.054 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:01:55.055 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:01:55.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:55 vm08.local ceph-mon[50506]: pgmap v3033: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:55.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:55 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:01:55.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:55 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:01:55.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:55 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:01:55.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:55 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:01:55.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:55 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:01:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:55 vm01.local ceph-mon[48597]: pgmap v3033: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:55 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:01:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:55 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:01:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:55 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:01:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:55 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:01:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:55 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:01:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:57 vm08.local ceph-mon[50506]: pgmap v3034: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:57 vm01.local ceph-mon[48597]: pgmap v3034: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:01:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:01:59 vm08.local ceph-mon[50506]: pgmap v3035: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:01:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:01:59 vm01.local ceph-mon[48597]: pgmap v3035: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:02:00.057 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:02:00.057 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:02:00.085 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:02:00.085 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:02:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:01 vm01.local ceph-mon[48597]: pgmap v3036: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:01.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:01 vm08.local ceph-mon[50506]: pgmap v3036: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:03 vm08.local ceph-mon[50506]: pgmap v3037: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:02:03.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:03 vm01.local ceph-mon[48597]: pgmap v3037: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:02:05.087 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:02:05.087 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:02:05.114 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:02:05.114 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:02:05.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:05 vm08.local ceph-mon[50506]: pgmap v3038: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:05.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:05 vm01.local ceph-mon[48597]: pgmap v3038: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:07 vm08.local ceph-mon[50506]: pgmap v3039: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:07 vm01.local ceph-mon[48597]: pgmap v3039: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:09 vm08.local ceph-mon[50506]: pgmap v3040: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:02:09.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:09 vm01.local ceph-mon[48597]: pgmap v3040: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:02:10.115 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:02:10.116 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:02:10.144 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:02:10.144 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:02:10.992 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:10 vm01.local ceph-mon[48597]: pgmap v3041: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:10.992 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:02:10.992 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:02:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:10 vm08.local ceph-mon[50506]: pgmap v3041: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:02:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:02:13.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:12 vm08.local ceph-mon[50506]: pgmap v3042: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:02:13.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:12 vm01.local ceph-mon[48597]: pgmap v3042: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:02:15.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:14 vm08.local ceph-mon[50506]: pgmap v3043: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:15.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:14 vm01.local ceph-mon[48597]: pgmap v3043: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:15.146 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:02:15.147 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:02:15.172 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:02:15.172 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:02:17.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:16 vm08.local ceph-mon[50506]: pgmap v3044: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:17.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:16 vm01.local ceph-mon[48597]: pgmap v3044: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:19.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:18 vm08.local ceph-mon[50506]: pgmap v3045: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:02:19.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:18 vm01.local ceph-mon[48597]: pgmap v3045: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:02:20.173 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:02:20.174 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:02:20.200 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:02:20.200 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:02:21.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:20 vm08.local ceph-mon[50506]: pgmap v3046: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:21.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:20 vm01.local ceph-mon[48597]: pgmap v3046: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:23.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:22 vm08.local ceph-mon[50506]: pgmap v3047: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:02:23.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:22 vm01.local ceph-mon[48597]: pgmap v3047: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:02:25.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:24 vm08.local ceph-mon[50506]: pgmap v3048: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:25.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:24 vm01.local ceph-mon[48597]: pgmap v3048: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:25.201 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:02:25.202 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:02:25.230 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:02:25.231 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:02:27.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:26 vm08.local ceph-mon[50506]: pgmap v3049: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:27.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:26 vm01.local ceph-mon[48597]: pgmap v3049: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:29.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:28 vm08.local ceph-mon[50506]: pgmap v3050: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:02:29.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:28 vm01.local ceph-mon[48597]: pgmap v3050: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:02:30.232 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:02:30.233 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:02:30.262 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:02:30.262 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:02:31.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:30 vm08.local ceph-mon[50506]: pgmap v3051: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:31.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:30 vm01.local ceph-mon[48597]: pgmap v3051: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:33.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:32 vm08.local ceph-mon[50506]: pgmap v3052: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:02:33.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:32 vm01.local ceph-mon[48597]: pgmap v3052: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:02:35.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:34 vm08.local ceph-mon[50506]: pgmap v3053: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:35.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:34 vm01.local ceph-mon[48597]: pgmap v3053: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:35.264 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:02:35.264 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:02:35.294 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:02:35.294 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:02:37.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:36 vm08.local ceph-mon[50506]: pgmap v3054: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:37.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:36 vm01.local ceph-mon[48597]: pgmap v3054: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:39.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:38 vm08.local ceph-mon[50506]: pgmap v3055: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:02:39.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:38 vm01.local ceph-mon[48597]: pgmap v3055: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:02:40.296 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:02:40.296 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:02:40.493 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:02:40.494 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:02:41.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:40 vm01.local ceph-mon[48597]: pgmap v3056: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:41.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:40 vm08.local ceph-mon[50506]: pgmap v3056: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:43.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:42 vm01.local ceph-mon[48597]: pgmap v3057: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:02:43.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:42 vm08.local ceph-mon[50506]: pgmap v3057: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:02:45.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:44 vm08.local ceph-mon[50506]: pgmap v3058: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:45.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:44 vm01.local ceph-mon[48597]: pgmap v3058: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:45.495 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:02:45.496 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:02:45.522 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:02:45.523 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:02:47.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:46 vm08.local ceph-mon[50506]: pgmap v3059: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:47.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:46 vm01.local ceph-mon[48597]: pgmap v3059: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:49.170 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:48 vm01.local ceph-mon[48597]: pgmap v3060: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:02:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:48 vm08.local ceph-mon[50506]: pgmap v3060: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:02:50.524 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:02:50.524 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:02:50.552 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:02:50.552 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:02:51.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:50 vm08.local ceph-mon[50506]: pgmap v3061: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:50 vm01.local ceph-mon[48597]: pgmap v3061: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:53 vm08.local ceph-mon[50506]: pgmap v3062: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:02:53.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:53 vm01.local ceph-mon[48597]: pgmap v3062: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:02:55.554 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:02:55.554 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:02:55.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:55 vm08.local ceph-mon[50506]: pgmap v3063: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:55.607 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:55 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:02:55.608 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:55 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:02:55.608 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:55 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:02:55.608 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:55 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:02:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:55 vm01.local ceph-mon[48597]: pgmap v3063: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:55.648 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:55 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:02:55.648 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:55 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:02:55.648 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:55 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:02:55.648 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:55 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:02:55.738 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:02:55.738 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:02:56.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:56 vm08.local ceph-mon[50506]: pgmap v3064: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:56.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:56 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:02:56.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:56 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:02:56.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:56 vm01.local ceph-mon[48597]: pgmap v3064: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:56.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:56 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:02:56.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:56 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:02:58.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:02:58 vm08.local ceph-mon[50506]: pgmap v3065: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:02:58.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:02:58 vm01.local ceph-mon[48597]: pgmap v3065: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:00.739 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:03:00.740 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:03:00.766 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:03:00.767 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:03:00.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:00 vm01.local ceph-mon[48597]: pgmap v3066: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T12:03:01.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:00 vm08.local ceph-mon[50506]: pgmap v3066: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T12:03:03.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:02 vm08.local ceph-mon[50506]: pgmap v3067: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:03.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:02 vm01.local ceph-mon[48597]: pgmap v3067: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:05.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:04 vm08.local ceph-mon[50506]: pgmap v3068: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T12:03:05.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:04 vm01.local ceph-mon[48597]: pgmap v3068: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T12:03:05.768 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:03:05.768 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:03:05.797 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:03:05.798 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:03:07.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:06 vm08.local ceph-mon[50506]: pgmap v3069: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T12:03:07.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:06 vm01.local ceph-mon[48597]: pgmap v3069: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T12:03:09.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:08 vm08.local ceph-mon[50506]: pgmap v3070: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:09.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:08 vm01.local ceph-mon[48597]: pgmap v3070: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:10.799 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:03:10.799 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:03:10.826 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:03:10.826 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:03:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:11 vm08.local ceph-mon[50506]: pgmap v3071: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:03:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:03:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:11 vm01.local ceph-mon[48597]: pgmap v3071: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:03:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:03:13.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:13 vm01.local ceph-mon[48597]: pgmap v3072: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:03:13.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:13 vm08.local ceph-mon[50506]: pgmap v3072: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:03:15.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:15 vm01.local ceph-mon[48597]: pgmap v3073: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:15.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:15 vm08.local ceph-mon[50506]: pgmap v3073: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:15.828 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:03:15.829 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:03:15.857 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:03:15.858 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:03:17.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:17 vm01.local ceph-mon[48597]: pgmap v3074: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:17 vm08.local ceph-mon[50506]: pgmap v3074: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:19.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:19 vm08.local ceph-mon[50506]: pgmap v3075: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:03:19.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:19 vm01.local ceph-mon[48597]: pgmap v3075: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:03:20.860 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:03:20.860 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:03:20.887 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:03:20.888 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:03:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:21 vm01.local ceph-mon[48597]: pgmap v3076: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:21.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:21 vm08.local ceph-mon[50506]: pgmap v3076: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:23 vm08.local ceph-mon[50506]: pgmap v3077: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:03:23.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:23 vm01.local ceph-mon[48597]: pgmap v3077: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:03:25.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:25 vm08.local ceph-mon[50506]: pgmap v3078: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:25.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:25 vm01.local ceph-mon[48597]: pgmap v3078: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:25.889 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:03:25.890 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:03:25.916 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:03:25.916 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:03:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:27 vm08.local ceph-mon[50506]: pgmap v3079: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:27 vm01.local ceph-mon[48597]: pgmap v3079: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:28.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:28 vm08.local ceph-mon[50506]: pgmap v3080: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:03:28.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:28 vm01.local ceph-mon[48597]: pgmap v3080: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:03:30.918 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:03:30.918 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:03:30.945 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:03:30.946 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:03:31.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:30 vm08.local ceph-mon[50506]: pgmap v3081: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:31.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:30 vm01.local ceph-mon[48597]: pgmap v3081: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:33 vm08.local ceph-mon[50506]: pgmap v3082: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:03:33.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:33 vm01.local ceph-mon[48597]: pgmap v3082: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:03:35.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:35 vm08.local ceph-mon[50506]: pgmap v3083: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:35.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:35 vm01.local ceph-mon[48597]: pgmap v3083: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:35.947 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:03:35.948 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:03:35.974 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:03:35.975 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:03:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:37 vm08.local ceph-mon[50506]: pgmap v3084: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:37 vm01.local ceph-mon[48597]: pgmap v3084: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:39 vm08.local ceph-mon[50506]: pgmap v3085: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:03:39.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:39 vm01.local ceph-mon[48597]: pgmap v3085: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:03:40.976 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:03:40.977 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:03:41.004 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:03:41.005 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:03:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:41 vm01.local ceph-mon[48597]: pgmap v3086: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:41.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:41 vm08.local ceph-mon[50506]: pgmap v3086: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:43 vm08.local ceph-mon[50506]: pgmap v3087: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:03:43.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:43 vm01.local ceph-mon[48597]: pgmap v3087: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:03:45.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:45 vm08.local ceph-mon[50506]: pgmap v3088: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:45.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:45 vm01.local ceph-mon[48597]: pgmap v3088: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:46.006 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:03:46.007 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:03:46.032 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:03:46.033 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:03:47.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:47 vm08.local ceph-mon[50506]: pgmap v3089: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:47.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:47 vm01.local ceph-mon[48597]: pgmap v3089: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:49.420 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:49 vm01.local ceph-mon[48597]: pgmap v3090: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:03:49.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:49 vm08.local ceph-mon[50506]: pgmap v3090: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:03:51.034 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:03:51.035 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:03:51.061 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:03:51.062 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:03:51.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:51 vm08.local ceph-mon[50506]: pgmap v3091: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:51 vm01.local ceph-mon[48597]: pgmap v3091: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:53.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:53 vm08.local ceph-mon[50506]: pgmap v3092: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:03:53.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:53 vm01.local ceph-mon[48597]: pgmap v3092: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:03:55.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:55 vm08.local ceph-mon[50506]: pgmap v3093: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:55 vm01.local ceph-mon[48597]: pgmap v3093: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:56.064 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:03:56.064 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:03:56.092 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:03:56.093 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:03:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:56 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:03:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:56 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:03:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:56 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:03:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:56 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:03:56.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:56 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:03:56.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:56 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:03:56.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:56 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:03:56.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:56 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:03:56.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:56 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:03:56.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:56 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:03:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:57 vm08.local ceph-mon[50506]: pgmap v3094: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:57 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:03:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:57 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:03:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:57 vm01.local ceph-mon[48597]: pgmap v3094: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:03:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:57 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:03:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:57 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:03:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:03:59 vm08.local ceph-mon[50506]: pgmap v3095: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:03:59.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:03:59 vm01.local ceph-mon[48597]: pgmap v3095: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:04:01.094 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:04:01.095 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:04:01.120 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:04:01.121 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:04:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:01 vm01.local ceph-mon[48597]: pgmap v3096: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:01.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:01 vm08.local ceph-mon[50506]: pgmap v3096: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:03 vm08.local ceph-mon[50506]: pgmap v3097: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:04:03.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:03 vm01.local ceph-mon[48597]: pgmap v3097: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:04:05.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:05 vm08.local ceph-mon[50506]: pgmap v3098: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:05.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:05 vm01.local ceph-mon[48597]: pgmap v3098: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:06.122 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:04:06.123 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:04:06.201 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:04:06.201 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:04:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:07 vm08.local ceph-mon[50506]: pgmap v3099: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:07 vm01.local ceph-mon[48597]: pgmap v3099: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:09 vm08.local ceph-mon[50506]: pgmap v3100: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:04:09.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:09 vm01.local ceph-mon[48597]: pgmap v3100: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:04:11.202 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:04:11.203 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:04:11.241 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:04:11.242 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:04:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:11 vm01.local ceph-mon[48597]: pgmap v3101: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:04:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:04:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:11 vm08.local ceph-mon[50506]: pgmap v3101: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:04:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:04:13.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:13 vm08.local ceph-mon[50506]: pgmap v3102: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:04:13.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:13 vm01.local ceph-mon[48597]: pgmap v3102: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:04:15.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:15 vm08.local ceph-mon[50506]: pgmap v3103: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:15.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:15 vm01.local ceph-mon[48597]: pgmap v3103: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:16.243 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:04:16.243 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:04:16.269 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:04:16.270 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:04:16.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:16 vm08.local ceph-mon[50506]: pgmap v3104: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:16.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:16 vm01.local ceph-mon[48597]: pgmap v3104: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:18.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:18 vm08.local ceph-mon[50506]: pgmap v3105: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:04:18.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:18 vm01.local ceph-mon[48597]: pgmap v3105: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:04:21.271 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:04:21.272 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:04:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:20 vm08.local ceph-mon[50506]: pgmap v3106: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:21.371 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:20 vm01.local ceph-mon[48597]: pgmap v3106: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:21.536 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:04:21.537 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:04:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:23 vm08.local ceph-mon[50506]: pgmap v3107: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:04:23.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:23 vm01.local ceph-mon[48597]: pgmap v3107: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:04:25.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:25 vm08.local ceph-mon[50506]: pgmap v3108: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:25.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:25 vm01.local ceph-mon[48597]: pgmap v3108: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:26.538 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:04:26.539 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:04:26.565 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:04:26.565 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:04:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:27 vm08.local ceph-mon[50506]: pgmap v3109: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:27 vm01.local ceph-mon[48597]: pgmap v3109: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:29 vm08.local ceph-mon[50506]: pgmap v3110: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:04:29.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:29 vm01.local ceph-mon[48597]: pgmap v3110: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:04:31.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:31 vm01.local ceph-mon[48597]: pgmap v3111: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:31.567 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:04:31.567 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:04:31.593 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:04:31.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:31 vm08.local ceph-mon[50506]: pgmap v3111: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:31.594 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:04:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:33 vm08.local ceph-mon[50506]: pgmap v3112: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:04:33.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:33 vm01.local ceph-mon[48597]: pgmap v3112: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:04:35.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:35 vm08.local ceph-mon[50506]: pgmap v3113: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:35.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:35 vm01.local ceph-mon[48597]: pgmap v3113: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:36.595 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:04:36.596 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:04:36.621 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:04:36.622 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:04:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:37 vm08.local ceph-mon[50506]: pgmap v3114: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:37 vm01.local ceph-mon[48597]: pgmap v3114: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:38.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:38 vm08.local ceph-mon[50506]: pgmap v3115: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:04:38.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:38 vm01.local ceph-mon[48597]: pgmap v3115: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:04:41.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:40 vm08.local ceph-mon[50506]: pgmap v3116: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:40 vm01.local ceph-mon[48597]: pgmap v3116: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:41.623 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:04:41.624 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:04:41.650 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:04:41.651 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:04:43.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:42 vm08.local ceph-mon[50506]: pgmap v3117: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:04:43.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:42 vm01.local ceph-mon[48597]: pgmap v3117: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:04:45.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:44 vm08.local ceph-mon[50506]: pgmap v3118: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:45.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:44 vm01.local ceph-mon[48597]: pgmap v3118: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:46.652 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:04:46.653 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:04:46.715 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:04:46.715 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:04:47.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:46 vm08.local ceph-mon[50506]: pgmap v3119: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:47.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:46 vm01.local ceph-mon[48597]: pgmap v3119: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:48 vm08.local ceph-mon[50506]: pgmap v3120: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:04:49.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:48 vm01.local ceph-mon[48597]: pgmap v3120: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:04:51.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:50 vm08.local ceph-mon[50506]: pgmap v3121: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:50 vm01.local ceph-mon[48597]: pgmap v3121: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:51.717 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:04:51.717 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:04:51.743 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:04:51.744 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:04:53.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:52 vm08.local ceph-mon[50506]: pgmap v3122: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:04:53.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:52 vm01.local ceph-mon[48597]: pgmap v3122: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:04:55.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:54 vm08.local ceph-mon[50506]: pgmap v3123: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:54 vm01.local ceph-mon[48597]: pgmap v3123: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:56.745 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:04:56.746 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:04:56.782 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:04:56.783 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:04:57.230 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:56 vm08.local ceph-mon[50506]: pgmap v3124: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:57.230 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:56 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:04:57.230 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:56 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:04:57.230 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:56 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:04:57.354 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:56 vm01.local ceph-mon[48597]: pgmap v3124: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:04:57.354 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:56 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:04:57.354 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:56 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:04:57.354 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:56 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:04:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:57 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:04:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:57 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:04:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:57 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:04:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:57 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:04:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:57 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:04:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:57 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:04:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:57 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:04:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:57 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:04:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:57 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:04:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:57 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:04:59.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:04:59 vm08.local ceph-mon[50506]: pgmap v3125: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:04:59.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:04:59 vm01.local ceph-mon[48597]: pgmap v3125: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:05:01.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:01 vm08.local ceph-mon[50506]: pgmap v3126: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:01 vm01.local ceph-mon[48597]: pgmap v3126: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:01.784 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:05:01.785 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:05:01.811 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:05:01.812 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:05:03.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:03 vm08.local ceph-mon[50506]: pgmap v3127: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:05:03.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:03 vm01.local ceph-mon[48597]: pgmap v3127: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:05:05.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:05 vm08.local ceph-mon[50506]: pgmap v3128: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:05.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:05 vm01.local ceph-mon[48597]: pgmap v3128: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:06.813 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:05:06.814 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:05:06.842 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:05:06.842 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:05:07.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:07 vm08.local ceph-mon[50506]: pgmap v3129: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:07.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:07 vm01.local ceph-mon[48597]: pgmap v3129: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:09.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:09 vm08.local ceph-mon[50506]: pgmap v3130: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:05:09.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:09 vm01.local ceph-mon[48597]: pgmap v3130: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:05:11.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:11 vm01.local ceph-mon[48597]: pgmap v3131: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:11.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:05:11.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:05:11.844 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:05:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:11 vm08.local ceph-mon[50506]: pgmap v3131: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:05:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:05:11.844 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:05:11.879 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:05:11.880 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:05:12.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:12 vm08.local ceph-mon[50506]: pgmap v3132: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:05:12.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:12 vm01.local ceph-mon[48597]: pgmap v3132: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:05:14.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:14 vm08.local ceph-mon[50506]: pgmap v3133: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:14 vm01.local ceph-mon[48597]: pgmap v3133: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:16.881 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:05:16.882 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:05:16.907 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:05:16.907 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:05:17.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:16 vm08.local ceph-mon[50506]: pgmap v3134: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:17.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:16 vm01.local ceph-mon[48597]: pgmap v3134: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:19.316 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:18 vm01.local ceph-mon[48597]: pgmap v3135: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:05:19.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:18 vm08.local ceph-mon[50506]: pgmap v3135: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:05:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:20 vm08.local ceph-mon[50506]: pgmap v3136: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:20 vm01.local ceph-mon[48597]: pgmap v3136: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:21.909 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:05:21.910 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:05:21.936 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:05:21.936 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:05:23.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:22 vm08.local ceph-mon[50506]: pgmap v3137: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:05:23.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:22 vm01.local ceph-mon[48597]: pgmap v3137: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:05:25.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:25 vm08.local ceph-mon[50506]: pgmap v3138: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:25.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:25 vm01.local ceph-mon[48597]: pgmap v3138: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:26.938 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:05:26.938 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:05:26.966 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:05:26.966 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:05:27.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:27 vm08.local ceph-mon[50506]: pgmap v3139: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:27.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:27 vm01.local ceph-mon[48597]: pgmap v3139: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:29.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:29 vm08.local ceph-mon[50506]: pgmap v3140: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:05:29.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:29 vm01.local ceph-mon[48597]: pgmap v3140: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:05:31.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:31 vm08.local ceph-mon[50506]: pgmap v3141: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:31.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:31 vm01.local ceph-mon[48597]: pgmap v3141: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:31.968 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:05:31.968 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:05:31.994 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:05:31.994 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:05:33.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:33 vm08.local ceph-mon[50506]: pgmap v3142: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:05:33.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:33 vm01.local ceph-mon[48597]: pgmap v3142: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:05:35.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:35 vm08.local ceph-mon[50506]: pgmap v3143: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:35.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:35 vm01.local ceph-mon[48597]: pgmap v3143: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:36.996 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:05:36.996 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:05:37.023 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:05:37.023 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:05:37.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:37 vm08.local ceph-mon[50506]: pgmap v3144: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:37.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:37 vm01.local ceph-mon[48597]: pgmap v3144: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:39.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:39 vm08.local ceph-mon[50506]: pgmap v3145: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:05:39.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:39 vm01.local ceph-mon[48597]: pgmap v3145: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:05:41.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:41 vm08.local ceph-mon[50506]: pgmap v3146: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:41 vm01.local ceph-mon[48597]: pgmap v3146: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:42.025 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:05:42.026 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:05:42.066 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:05:42.067 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:05:43.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:43 vm08.local ceph-mon[50506]: pgmap v3147: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:05:43.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:43 vm01.local ceph-mon[48597]: pgmap v3147: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:05:45.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:45 vm08.local ceph-mon[50506]: pgmap v3148: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:45.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:45 vm01.local ceph-mon[48597]: pgmap v3148: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:47.068 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:05:47.069 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:05:47.096 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:05:47.097 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:05:47.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:47 vm08.local ceph-mon[50506]: pgmap v3149: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:47.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:47 vm01.local ceph-mon[48597]: pgmap v3149: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:49 vm08.local ceph-mon[50506]: pgmap v3150: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:05:49.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:49 vm01.local ceph-mon[48597]: pgmap v3150: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:05:51.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:51 vm08.local ceph-mon[50506]: pgmap v3151: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:51 vm01.local ceph-mon[48597]: pgmap v3151: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:52.098 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:05:52.099 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:05:52.128 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:05:52.129 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:05:53.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:53 vm08.local ceph-mon[50506]: pgmap v3152: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:05:53.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:53 vm01.local ceph-mon[48597]: pgmap v3152: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:05:55.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:55 vm08.local ceph-mon[50506]: pgmap v3153: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:55 vm01.local ceph-mon[48597]: pgmap v3153: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:57.130 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:05:57.130 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:05:57.157 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:05:57.158 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:05:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:57 vm08.local ceph-mon[50506]: pgmap v3154: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:57.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:57 vm01.local ceph-mon[48597]: pgmap v3154: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:05:58.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:58 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:05:58.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:58 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:05:58.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:58 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:05:58.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:58 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:05:58.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:58 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:05:58.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:58 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:05:58.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:58 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:05:58.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:58 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:05:58.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:58 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:05:58.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:58 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:05:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:05:59 vm08.local ceph-mon[50506]: pgmap v3155: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:05:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:05:59 vm01.local ceph-mon[48597]: pgmap v3155: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:06:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:01 vm01.local ceph-mon[48597]: pgmap v3156: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:01.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:01 vm08.local ceph-mon[50506]: pgmap v3156: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:02.159 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:06:02.183 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:06:02.296 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:06:02.297 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:06:02.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:02 vm01.local ceph-mon[48597]: pgmap v3157: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:06:02.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:02 vm08.local ceph-mon[50506]: pgmap v3157: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:06:04.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:04 vm01.local ceph-mon[48597]: pgmap v3158: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:04.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:04 vm08.local ceph-mon[50506]: pgmap v3158: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:07.298 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:06:07.299 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:06:07.326 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:06:07.326 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:06:07.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:07 vm08.local ceph-mon[50506]: pgmap v3159: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:07.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:07 vm01.local ceph-mon[48597]: pgmap v3159: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:09.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:09 vm01.local ceph-mon[48597]: pgmap v3160: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:06:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:09 vm08.local ceph-mon[50506]: pgmap v3160: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:06:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:11 vm01.local ceph-mon[48597]: pgmap v3161: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:06:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:06:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:11 vm08.local ceph-mon[50506]: pgmap v3161: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:06:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:06:12.328 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:06:12.329 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:06:12.355 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:06:12.356 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:06:13.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:13 vm01.local ceph-mon[48597]: pgmap v3162: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:06:13.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:13 vm08.local ceph-mon[50506]: pgmap v3162: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:06:15.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:15 vm08.local ceph-mon[50506]: pgmap v3163: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:15.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:15 vm01.local ceph-mon[48597]: pgmap v3163: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:17.357 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:06:17.358 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:06:17.386 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:06:17.386 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:06:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:17 vm08.local ceph-mon[50506]: pgmap v3164: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:17 vm01.local ceph-mon[48597]: pgmap v3164: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:19.420 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:19 vm01.local ceph-mon[48597]: pgmap v3165: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:06:19.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:19 vm08.local ceph-mon[50506]: pgmap v3165: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:06:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:21 vm01.local ceph-mon[48597]: pgmap v3166: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:21.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:21 vm08.local ceph-mon[50506]: pgmap v3166: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:22.388 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:06:22.389 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:06:22.415 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:06:22.415 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:06:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:23 vm08.local ceph-mon[50506]: pgmap v3167: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:06:23.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:23 vm01.local ceph-mon[48597]: pgmap v3167: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:06:25.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:25 vm08.local ceph-mon[50506]: pgmap v3168: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:25.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:25 vm01.local ceph-mon[48597]: pgmap v3168: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:27.417 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:06:27.417 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:06:27.444 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:06:27.444 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:06:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:27 vm08.local ceph-mon[50506]: pgmap v3169: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:27 vm01.local ceph-mon[48597]: pgmap v3169: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:29 vm08.local ceph-mon[50506]: pgmap v3170: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:06:29.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:29 vm01.local ceph-mon[48597]: pgmap v3170: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:06:31.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:31 vm08.local ceph-mon[50506]: pgmap v3171: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:31.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:31 vm01.local ceph-mon[48597]: pgmap v3171: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:32.446 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:06:32.446 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:06:32.472 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:06:32.473 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:06:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:33 vm08.local ceph-mon[50506]: pgmap v3172: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:06:33.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:33 vm01.local ceph-mon[48597]: pgmap v3172: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:06:35.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:35 vm08.local ceph-mon[50506]: pgmap v3173: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:35.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:35 vm01.local ceph-mon[48597]: pgmap v3173: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:36.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:36 vm08.local ceph-mon[50506]: pgmap v3174: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:36.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:36 vm01.local ceph-mon[48597]: pgmap v3174: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:37.474 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:06:37.475 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:06:37.500 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:06:37.501 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:06:38.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:38 vm08.local ceph-mon[50506]: pgmap v3175: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:06:38.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:38 vm01.local ceph-mon[48597]: pgmap v3175: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:06:41.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:41 vm08.local ceph-mon[50506]: pgmap v3176: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:41 vm01.local ceph-mon[48597]: pgmap v3176: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:42.502 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:06:42.503 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:06:42.528 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:06:42.529 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:06:43.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:43 vm08.local ceph-mon[50506]: pgmap v3177: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:06:43.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:43 vm01.local ceph-mon[48597]: pgmap v3177: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:06:45.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:45 vm08.local ceph-mon[50506]: pgmap v3178: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:45.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:45 vm01.local ceph-mon[48597]: pgmap v3178: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:47.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:47 vm08.local ceph-mon[50506]: pgmap v3179: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:47.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:47 vm01.local ceph-mon[48597]: pgmap v3179: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:47.530 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:06:47.530 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:06:47.557 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:06:47.557 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:06:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:49 vm08.local ceph-mon[50506]: pgmap v3180: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:06:49.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:49 vm01.local ceph-mon[48597]: pgmap v3180: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:06:51.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:51 vm08.local ceph-mon[50506]: pgmap v3181: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:51 vm01.local ceph-mon[48597]: pgmap v3181: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:52.558 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:06:52.559 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:06:52.585 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:06:52.585 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:06:53.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:53 vm08.local ceph-mon[50506]: pgmap v3182: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:06:53.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:53 vm01.local ceph-mon[48597]: pgmap v3182: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:06:55.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:55 vm08.local ceph-mon[50506]: pgmap v3183: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:55 vm01.local ceph-mon[48597]: pgmap v3183: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:57.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:57 vm08.local ceph-mon[50506]: pgmap v3184: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:57.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:57 vm01.local ceph-mon[48597]: pgmap v3184: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:06:57.586 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:06:57.587 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:06:57.617 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:06:57.618 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:06:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:58 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:06:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:58 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:06:58.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:58 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:06:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:58 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:06:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:58 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:06:58.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:58 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:06:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:59 vm08.local ceph-mon[50506]: pgmap v3185: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:06:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:59 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:06:59.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:06:59 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:06:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:59 vm01.local ceph-mon[48597]: pgmap v3185: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:06:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:59 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:06:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:06:59 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:07:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:01 vm01.local ceph-mon[48597]: pgmap v3186: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:01.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:01 vm08.local ceph-mon[50506]: pgmap v3186: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:02.619 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:07:02.619 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:07:02.646 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:07:02.647 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:07:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:03 vm08.local ceph-mon[50506]: pgmap v3187: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:07:03.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:03 vm01.local ceph-mon[48597]: pgmap v3187: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:07:04.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:04 vm01.local ceph-mon[48597]: pgmap v3188: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:04.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:04 vm08.local ceph-mon[50506]: pgmap v3188: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:07.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:07 vm08.local ceph-mon[50506]: pgmap v3189: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:07.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:07 vm01.local ceph-mon[48597]: pgmap v3189: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:07.648 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:07:07.649 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:07:07.676 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:07:07.676 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:07:09.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:09 vm01.local ceph-mon[48597]: pgmap v3190: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:07:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:09 vm08.local ceph-mon[50506]: pgmap v3190: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:07:10.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:07:10.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:07:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:11 vm08.local ceph-mon[50506]: pgmap v3191: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:07:11.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:11 vm01.local ceph-mon[48597]: pgmap v3191: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:11.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:07:12.678 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:07:12.678 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:07:12.704 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:07:12.704 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:07:13.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:13 vm01.local ceph-mon[48597]: pgmap v3192: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:07:13.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:13 vm08.local ceph-mon[50506]: pgmap v3192: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:07:14.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:14 vm08.local ceph-mon[50506]: pgmap v3193: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:14 vm01.local ceph-mon[48597]: pgmap v3193: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:17 vm08.local ceph-mon[50506]: pgmap v3194: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:17 vm01.local ceph-mon[48597]: pgmap v3194: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:17.706 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:07:17.706 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:07:17.733 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:07:17.733 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:07:19.420 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:19 vm01.local ceph-mon[48597]: pgmap v3195: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:07:19.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:19 vm08.local ceph-mon[50506]: pgmap v3195: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:07:21.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:21 vm08.local ceph-mon[50506]: pgmap v3196: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:21.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:21 vm01.local ceph-mon[48597]: pgmap v3196: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:22.735 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:07:22.735 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:07:22.761 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:07:22.761 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:07:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:23 vm08.local ceph-mon[50506]: pgmap v3197: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:07:23.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:23 vm01.local ceph-mon[48597]: pgmap v3197: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:07:25.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:25 vm08.local ceph-mon[50506]: pgmap v3198: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:25.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:25 vm01.local ceph-mon[48597]: pgmap v3198: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:27 vm08.local ceph-mon[50506]: pgmap v3199: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:27 vm01.local ceph-mon[48597]: pgmap v3199: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:27.763 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:07:27.763 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:07:27.933 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:07:27.934 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:07:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:29 vm08.local ceph-mon[50506]: pgmap v3200: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:07:29.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:29 vm01.local ceph-mon[48597]: pgmap v3200: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:07:31.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:31 vm01.local ceph-mon[48597]: pgmap v3201: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:31.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:31 vm08.local ceph-mon[50506]: pgmap v3201: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:32.935 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:07:32.935 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:07:32.960 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:07:32.961 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:07:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:33 vm08.local ceph-mon[50506]: pgmap v3202: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:07:33.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:33 vm01.local ceph-mon[48597]: pgmap v3202: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:07:35.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:35 vm01.local ceph-mon[48597]: pgmap v3203: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:35.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:35 vm08.local ceph-mon[50506]: pgmap v3203: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:36.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:36 vm08.local ceph-mon[50506]: pgmap v3204: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:36.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:36 vm01.local ceph-mon[48597]: pgmap v3204: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:37.962 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:07:37.962 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:07:38.000 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:07:38.000 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:07:38.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:38 vm08.local ceph-mon[50506]: pgmap v3205: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:07:38.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:38 vm01.local ceph-mon[48597]: pgmap v3205: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:07:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:41 vm01.local ceph-mon[48597]: pgmap v3206: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:41.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:41 vm08.local ceph-mon[50506]: pgmap v3206: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:43.001 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:07:43.002 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:07:43.083 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:07:43.083 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:07:43.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:43 vm08.local ceph-mon[50506]: pgmap v3207: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:07:43.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:43 vm01.local ceph-mon[48597]: pgmap v3207: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:07:45.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:45 vm08.local ceph-mon[50506]: pgmap v3208: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:45.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:45 vm01.local ceph-mon[48597]: pgmap v3208: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:46.745 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:46 vm08.local ceph-mon[50506]: pgmap v3209: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:46.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:46 vm01.local ceph-mon[48597]: pgmap v3209: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:48.084 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:07:48.085 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:07:48.112 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:07:48.112 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:07:48.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:48 vm08.local ceph-mon[50506]: pgmap v3210: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:07:48.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:48 vm01.local ceph-mon[48597]: pgmap v3210: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:07:51.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:51 vm08.local ceph-mon[50506]: pgmap v3211: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:51.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:51 vm01.local ceph-mon[48597]: pgmap v3211: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:52.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:52 vm08.local ceph-mon[50506]: pgmap v3212: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:07:52.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:52 vm01.local ceph-mon[48597]: pgmap v3212: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:07:53.113 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:07:53.114 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:07:53.140 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:07:53.140 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:07:54.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:54 vm08.local ceph-mon[50506]: pgmap v3213: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:54.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:54 vm01.local ceph-mon[48597]: pgmap v3213: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:56.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:56 vm08.local ceph-mon[50506]: pgmap v3214: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:56.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:56 vm01.local ceph-mon[48597]: pgmap v3214: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:07:58.142 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:07:58.142 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:07:58.169 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:07:58.170 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:07:58.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:58 vm08.local ceph-mon[50506]: pgmap v3215: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:07:58.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:58 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:07:58.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:58 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:07:58.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:58 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:07:58.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:58 vm01.local ceph-mon[48597]: pgmap v3215: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:07:58.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:58 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:07:58.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:58 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:07:58.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:58 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:07:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:59 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:07:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:07:59 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:07:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:59 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:07:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:07:59 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:08:00.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:00 vm08.local ceph-mon[50506]: pgmap v3216: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:00.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:00 vm01.local ceph-mon[48597]: pgmap v3216: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:02.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:02 vm08.local ceph-mon[50506]: pgmap v3217: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:08:02.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:02 vm01.local ceph-mon[48597]: pgmap v3217: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:08:03.171 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:08:03.172 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:08:03.198 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:08:03.198 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:08:04.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:04 vm08.local ceph-mon[50506]: pgmap v3218: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:04.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:04 vm01.local ceph-mon[48597]: pgmap v3218: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:06.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:06 vm08.local ceph-mon[50506]: pgmap v3219: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:06.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:06 vm01.local ceph-mon[48597]: pgmap v3219: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:08.199 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:08:08.200 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:08:08.225 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:08:08.226 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:08:08.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:08 vm08.local ceph-mon[50506]: pgmap v3220: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:08:08.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:08 vm01.local ceph-mon[48597]: pgmap v3220: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:08:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:10 vm08.local ceph-mon[50506]: pgmap v3221: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:08:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:10 vm01.local ceph-mon[48597]: pgmap v3221: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:08:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:08:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:08:12.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:12 vm08.local ceph-mon[50506]: pgmap v3222: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:08:12.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:12 vm01.local ceph-mon[48597]: pgmap v3222: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:08:13.228 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:08:13.228 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:08:13.253 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:08:13.254 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:08:14.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:14 vm08.local ceph-mon[50506]: pgmap v3223: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:14 vm01.local ceph-mon[48597]: pgmap v3223: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:16.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:16 vm08.local ceph-mon[50506]: pgmap v3224: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:16.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:16 vm01.local ceph-mon[48597]: pgmap v3224: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:18.255 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:08:18.255 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:08:18.291 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:08:18.291 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:08:18.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:18 vm08.local ceph-mon[50506]: pgmap v3225: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:08:18.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:18 vm01.local ceph-mon[48597]: pgmap v3225: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:08:20.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:20 vm08.local ceph-mon[50506]: pgmap v3226: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:20.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:20 vm01.local ceph-mon[48597]: pgmap v3226: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:22.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:22 vm08.local ceph-mon[50506]: pgmap v3227: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:08:22.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:22 vm01.local ceph-mon[48597]: pgmap v3227: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:08:23.292 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:08:23.293 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:08:23.320 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:08:23.320 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:08:24.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:24 vm08.local ceph-mon[50506]: pgmap v3228: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:24.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:24 vm01.local ceph-mon[48597]: pgmap v3228: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:26.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:26 vm08.local ceph-mon[50506]: pgmap v3229: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:26.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:26 vm01.local ceph-mon[48597]: pgmap v3229: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:28.322 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:08:28.322 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:08:28.351 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:08:28.351 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:08:28.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:28 vm08.local ceph-mon[50506]: pgmap v3230: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:08:28.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:28 vm01.local ceph-mon[48597]: pgmap v3230: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:08:30.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:30 vm08.local ceph-mon[50506]: pgmap v3231: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:30.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:30 vm01.local ceph-mon[48597]: pgmap v3231: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:32.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:32 vm08.local ceph-mon[50506]: pgmap v3232: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:08:32.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:32 vm01.local ceph-mon[48597]: pgmap v3232: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:08:33.353 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:08:33.354 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:08:33.379 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:08:33.379 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:08:34.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:34 vm08.local ceph-mon[50506]: pgmap v3233: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:34.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:34 vm01.local ceph-mon[48597]: pgmap v3233: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:36.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:36 vm08.local ceph-mon[50506]: pgmap v3234: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:36.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:36 vm01.local ceph-mon[48597]: pgmap v3234: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:38.381 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:08:38.381 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:08:38.407 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:08:38.408 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:08:38.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:38 vm08.local ceph-mon[50506]: pgmap v3235: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:08:38.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:38 vm01.local ceph-mon[48597]: pgmap v3235: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:08:40.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:40 vm08.local ceph-mon[50506]: pgmap v3236: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:40.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:40 vm01.local ceph-mon[48597]: pgmap v3236: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:42.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:42 vm08.local ceph-mon[50506]: pgmap v3237: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:08:42.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:42 vm01.local ceph-mon[48597]: pgmap v3237: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:08:43.409 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:08:43.410 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:08:43.435 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:08:43.435 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:08:44.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:44 vm08.local ceph-mon[50506]: pgmap v3238: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:44.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:44 vm01.local ceph-mon[48597]: pgmap v3238: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:46.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:46 vm08.local ceph-mon[50506]: pgmap v3239: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:46.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:46 vm01.local ceph-mon[48597]: pgmap v3239: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:48.437 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:08:48.437 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:08:48.463 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:08:48.463 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:08:48.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:48 vm08.local ceph-mon[50506]: pgmap v3240: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:08:48.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:48 vm01.local ceph-mon[48597]: pgmap v3240: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:08:50.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:50 vm01.local ceph-mon[48597]: pgmap v3241: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:50.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:50 vm08.local ceph-mon[50506]: pgmap v3241: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:52.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:52 vm08.local ceph-mon[50506]: pgmap v3242: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:08:52.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:52 vm01.local ceph-mon[48597]: pgmap v3242: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:08:53.465 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:08:53.466 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:08:53.493 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:08:53.494 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:08:54.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:54 vm08.local ceph-mon[50506]: pgmap v3243: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:54.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:54 vm01.local ceph-mon[48597]: pgmap v3243: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:56.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:56 vm08.local ceph-mon[50506]: pgmap v3244: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:56.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:56 vm01.local ceph-mon[48597]: pgmap v3244: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:08:58.496 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:08:58.496 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:08:58.536 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:08:58.536 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:08:58.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:58 vm08.local ceph-mon[50506]: pgmap v3245: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:08:58.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:58 vm01.local ceph-mon[48597]: pgmap v3245: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:08:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:59 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:08:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:59 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:08:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:59 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:08:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:59 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:08:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:08:59 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:08:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:59 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:08:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:59 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:08:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:59 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:08:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:59 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:08:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:08:59 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:09:00.992 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:00 vm01.local ceph-mon[48597]: pgmap v3246: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:01.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:00 vm08.local ceph-mon[50506]: pgmap v3246: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:03.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:02 vm08.local ceph-mon[50506]: pgmap v3247: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:09:03.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:02 vm01.local ceph-mon[48597]: pgmap v3247: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:09:03.540 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:09:03.541 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:09:03.567 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:09:03.568 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:09:05.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:04 vm08.local ceph-mon[50506]: pgmap v3248: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:05.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:04 vm01.local ceph-mon[48597]: pgmap v3248: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:06.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:06 vm08.local ceph-mon[50506]: pgmap v3249: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:06.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:06 vm01.local ceph-mon[48597]: pgmap v3249: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:08.570 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:09:08.570 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:09:08.597 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:09:08.597 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:09:08.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:08 vm08.local ceph-mon[50506]: pgmap v3250: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:09:08.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:08 vm01.local ceph-mon[48597]: pgmap v3250: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:09:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:10 vm08.local ceph-mon[50506]: pgmap v3251: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:09:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:10 vm01.local ceph-mon[48597]: pgmap v3251: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:09:11.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:09:11.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:09:12.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:12 vm08.local ceph-mon[50506]: pgmap v3252: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:09:12.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:12 vm01.local ceph-mon[48597]: pgmap v3252: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:09:13.598 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:09:13.599 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:09:13.626 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:09:13.626 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:09:14.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:14 vm08.local ceph-mon[50506]: pgmap v3253: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:14 vm01.local ceph-mon[48597]: pgmap v3253: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:16.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:16 vm08.local ceph-mon[50506]: pgmap v3254: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:16.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:16 vm01.local ceph-mon[48597]: pgmap v3254: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:18.627 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:09:18.628 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:09:18.654 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:09:18.654 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:09:18.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:18 vm08.local ceph-mon[50506]: pgmap v3255: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:09:18.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:18 vm01.local ceph-mon[48597]: pgmap v3255: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:09:20.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:20 vm08.local ceph-mon[50506]: pgmap v3256: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:20.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:20 vm01.local ceph-mon[48597]: pgmap v3256: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:22.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:22 vm08.local ceph-mon[50506]: pgmap v3257: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:09:22.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:22 vm01.local ceph-mon[48597]: pgmap v3257: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:09:23.655 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:09:23.656 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:09:23.683 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:09:23.684 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:09:24.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:24 vm08.local ceph-mon[50506]: pgmap v3258: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:24.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:24 vm01.local ceph-mon[48597]: pgmap v3258: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:26.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:26 vm08.local ceph-mon[50506]: pgmap v3259: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:26.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:26 vm01.local ceph-mon[48597]: pgmap v3259: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:28.685 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:09:28.686 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:09:28.712 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:09:28.713 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:09:28.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:28 vm08.local ceph-mon[50506]: pgmap v3260: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:09:28.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:28 vm01.local ceph-mon[48597]: pgmap v3260: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:09:30.992 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:30 vm01.local ceph-mon[48597]: pgmap v3261: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:31.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:30 vm08.local ceph-mon[50506]: pgmap v3261: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:33.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:32 vm08.local ceph-mon[50506]: pgmap v3262: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:09:33.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:32 vm01.local ceph-mon[48597]: pgmap v3262: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:09:33.714 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:09:33.715 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:09:33.742 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:09:33.742 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:09:35.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:34 vm08.local ceph-mon[50506]: pgmap v3263: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:35.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:34 vm01.local ceph-mon[48597]: pgmap v3263: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:37.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:36 vm08.local ceph-mon[50506]: pgmap v3264: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:37.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:36 vm01.local ceph-mon[48597]: pgmap v3264: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:38.743 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:09:38.744 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:09:38.773 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:09:38.773 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:09:39.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:38 vm08.local ceph-mon[50506]: pgmap v3265: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:09:39.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:38 vm01.local ceph-mon[48597]: pgmap v3265: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:09:41.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:40 vm08.local ceph-mon[50506]: pgmap v3266: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:41.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:40 vm01.local ceph-mon[48597]: pgmap v3266: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:43.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:42 vm08.local ceph-mon[50506]: pgmap v3267: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:09:43.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:42 vm01.local ceph-mon[48597]: pgmap v3267: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:09:43.775 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:09:43.776 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:09:43.804 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:09:43.804 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:09:45.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:44 vm08.local ceph-mon[50506]: pgmap v3268: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:45.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:44 vm01.local ceph-mon[48597]: pgmap v3268: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:47.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:47 vm01.local ceph-mon[48597]: pgmap v3269: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:47.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:47 vm08.local ceph-mon[50506]: pgmap v3269: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:48.806 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:09:48.806 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:09:48.950 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:09:48.950 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:09:49.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:49 vm08.local ceph-mon[50506]: pgmap v3270: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:09:49.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:49 vm01.local ceph-mon[48597]: pgmap v3270: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:09:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:51 vm01.local ceph-mon[48597]: pgmap v3271: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:51.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:51 vm08.local ceph-mon[50506]: pgmap v3271: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:53.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:53 vm08.local ceph-mon[50506]: pgmap v3272: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:09:53.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:53 vm01.local ceph-mon[48597]: pgmap v3272: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:09:53.952 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:09:53.953 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:09:53.979 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:09:53.979 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:09:55.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:55 vm08.local ceph-mon[50506]: pgmap v3273: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:55.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:55 vm01.local ceph-mon[48597]: pgmap v3273: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:57.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:57 vm08.local ceph-mon[50506]: pgmap v3274: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:57.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:57 vm01.local ceph-mon[48597]: pgmap v3274: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:09:58.981 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:09:58.984 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:09:59.115 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:09:59.116 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:09:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:59 vm01.local ceph-mon[48597]: pgmap v3275: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:09:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:59 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:09:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:59 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:09:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:59 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:09:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:59 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:09:59.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:09:59 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:09:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:59 vm08.local ceph-mon[50506]: pgmap v3275: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:09:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:59 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:09:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:59 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:09:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:59 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:09:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:59 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:09:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:09:59 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:10:00.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:00 vm01.local ceph-mon[48597]: overall HEALTH_OK 2026-03-10T12:10:00.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:00 vm08.local ceph-mon[50506]: overall HEALTH_OK 2026-03-10T12:10:01.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:01 vm01.local ceph-mon[48597]: pgmap v3276: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:01.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:01 vm08.local ceph-mon[50506]: pgmap v3276: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:02.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:02 vm01.local ceph-mon[48597]: pgmap v3277: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:10:02.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:02 vm08.local ceph-mon[50506]: pgmap v3277: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:10:04.117 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:10:04.118 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:10:04.144 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:10:04.145 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:10:04.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:04 vm01.local ceph-mon[48597]: pgmap v3278: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:04.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:04 vm08.local ceph-mon[50506]: pgmap v3278: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:06.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:06 vm08.local ceph-mon[50506]: pgmap v3279: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:06.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:06 vm01.local ceph-mon[48597]: pgmap v3279: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:08.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:08 vm08.local ceph-mon[50506]: pgmap v3280: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:10:08.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:08 vm01.local ceph-mon[48597]: pgmap v3280: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:10:09.146 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:10:09.147 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:10:09.172 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:10:09.172 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:10:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:10 vm08.local ceph-mon[50506]: pgmap v3281: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:10:10.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:10:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:10 vm01.local ceph-mon[48597]: pgmap v3281: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:10:10.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:10:12.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:12 vm08.local ceph-mon[50506]: pgmap v3282: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:10:12.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:12 vm01.local ceph-mon[48597]: pgmap v3282: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:10:14.173 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:10:14.174 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:10:14.201 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:10:14.201 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:10:14.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:14 vm01.local ceph-mon[48597]: pgmap v3283: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:15.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:14 vm08.local ceph-mon[50506]: pgmap v3283: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:16.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:16 vm08.local ceph-mon[50506]: pgmap v3284: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:16.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:16 vm01.local ceph-mon[48597]: pgmap v3284: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:18.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:18 vm01.local ceph-mon[48597]: pgmap v3285: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:10:19.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:18 vm08.local ceph-mon[50506]: pgmap v3285: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:10:19.203 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:10:19.203 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:10:19.230 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:10:19.230 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:10:20.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:20 vm01.local ceph-mon[48597]: pgmap v3286: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:21.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:20 vm08.local ceph-mon[50506]: pgmap v3286: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:23.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:22 vm08.local ceph-mon[50506]: pgmap v3287: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:10:23.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:22 vm01.local ceph-mon[48597]: pgmap v3287: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:10:24.231 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:10:24.232 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:10:24.259 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:10:24.260 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:10:25.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:24 vm08.local ceph-mon[50506]: pgmap v3288: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:25.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:24 vm01.local ceph-mon[48597]: pgmap v3288: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:26.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:26 vm08.local ceph-mon[50506]: pgmap v3289: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:27.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:26 vm01.local ceph-mon[48597]: pgmap v3289: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:29.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:28 vm08.local ceph-mon[50506]: pgmap v3290: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:10:29.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:28 vm01.local ceph-mon[48597]: pgmap v3290: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:10:29.261 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:10:29.262 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:10:29.289 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:10:29.290 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:10:30.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:30 vm01.local ceph-mon[48597]: pgmap v3291: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:31.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:30 vm08.local ceph-mon[50506]: pgmap v3291: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:33.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:32 vm08.local ceph-mon[50506]: pgmap v3292: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:10:33.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:32 vm01.local ceph-mon[48597]: pgmap v3292: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:10:34.291 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:10:34.291 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:10:34.317 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:10:34.318 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:10:35.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:34 vm08.local ceph-mon[50506]: pgmap v3293: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:35.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:34 vm01.local ceph-mon[48597]: pgmap v3293: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:36.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:36 vm08.local ceph-mon[50506]: pgmap v3294: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:37.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:36 vm01.local ceph-mon[48597]: pgmap v3294: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:39.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:38 vm08.local ceph-mon[50506]: pgmap v3295: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:10:39.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:38 vm01.local ceph-mon[48597]: pgmap v3295: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:10:39.319 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:10:39.320 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:10:39.347 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:10:39.347 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:10:40.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:40 vm01.local ceph-mon[48597]: pgmap v3296: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:41.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:40 vm08.local ceph-mon[50506]: pgmap v3296: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:43.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:42 vm08.local ceph-mon[50506]: pgmap v3297: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:10:43.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:42 vm01.local ceph-mon[48597]: pgmap v3297: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:10:44.349 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:10:44.349 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:10:44.393 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:10:44.394 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:10:45.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:44 vm08.local ceph-mon[50506]: pgmap v3298: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:45.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:44 vm01.local ceph-mon[48597]: pgmap v3298: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:47.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:46 vm08.local ceph-mon[50506]: pgmap v3299: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:47.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:46 vm01.local ceph-mon[48597]: pgmap v3299: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:49 vm08.local ceph-mon[50506]: pgmap v3300: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:10:49.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:48 vm01.local ceph-mon[48597]: pgmap v3300: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:10:49.395 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:10:49.396 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:10:49.422 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:10:49.422 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:10:51.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:50 vm08.local ceph-mon[50506]: pgmap v3301: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:50 vm01.local ceph-mon[48597]: pgmap v3301: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:53.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:53 vm08.local ceph-mon[50506]: pgmap v3302: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:10:53.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:53 vm01.local ceph-mon[48597]: pgmap v3302: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:10:54.424 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:10:54.424 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:10:54.451 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:10:54.451 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:10:55.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:55 vm08.local ceph-mon[50506]: pgmap v3303: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:55 vm01.local ceph-mon[48597]: pgmap v3303: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:57.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:57 vm08.local ceph-mon[50506]: pgmap v3304: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:57.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:57 vm01.local ceph-mon[48597]: pgmap v3304: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:10:59.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:10:59 vm08.local ceph-mon[50506]: pgmap v3305: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:10:59.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:10:59 vm01.local ceph-mon[48597]: pgmap v3305: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:10:59.453 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:10:59.454 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:10:59.489 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:10:59.490 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:11:00.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:00 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:11:00.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:00 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:11:00.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:00 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:11:00.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:00 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:11:00.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:00 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:11:00.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:00 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:11:00.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:00 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:11:00.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:00 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:11:00.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:00 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:11:00.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:00 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:11:01.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:01 vm08.local ceph-mon[50506]: pgmap v3306: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:01 vm01.local ceph-mon[48597]: pgmap v3306: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:03.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:03 vm08.local ceph-mon[50506]: pgmap v3307: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:11:03.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:03 vm01.local ceph-mon[48597]: pgmap v3307: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:11:04.491 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:11:04.492 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:11:04.518 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:11:04.518 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:11:05.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:05 vm08.local ceph-mon[50506]: pgmap v3308: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:05.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:05 vm01.local ceph-mon[48597]: pgmap v3308: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:07.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:07 vm08.local ceph-mon[50506]: pgmap v3309: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:07.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:07 vm01.local ceph-mon[48597]: pgmap v3309: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:09.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:09 vm08.local ceph-mon[50506]: pgmap v3310: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:11:09.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:09 vm01.local ceph-mon[48597]: pgmap v3310: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:11:09.520 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:11:09.521 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:11:09.548 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:11:09.549 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:11:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:11 vm08.local ceph-mon[50506]: pgmap v3311: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:11:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:11:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:11 vm01.local ceph-mon[48597]: pgmap v3311: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:11:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:11:13.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:13 vm08.local ceph-mon[50506]: pgmap v3312: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:11:13.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:13 vm01.local ceph-mon[48597]: pgmap v3312: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:11:14.550 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:11:14.550 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:11:14.576 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:11:14.576 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:11:15.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:15 vm01.local ceph-mon[48597]: pgmap v3313: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:15.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:15 vm08.local ceph-mon[50506]: pgmap v3313: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:17.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:17 vm01.local ceph-mon[48597]: pgmap v3314: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:17 vm08.local ceph-mon[50506]: pgmap v3314: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:19.578 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:11:19.578 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:11:19.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:19 vm08.local ceph-mon[50506]: pgmap v3315: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:11:19.604 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:11:19.604 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:11:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:19 vm01.local ceph-mon[48597]: pgmap v3315: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:11:21.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:21 vm08.local ceph-mon[50506]: pgmap v3316: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:21.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:21 vm01.local ceph-mon[48597]: pgmap v3316: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:23 vm08.local ceph-mon[50506]: pgmap v3317: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:11:23.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:23 vm01.local ceph-mon[48597]: pgmap v3317: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:11:24.605 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:11:24.606 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:11:24.633 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:11:24.633 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:11:25.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:25 vm08.local ceph-mon[50506]: pgmap v3318: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:25.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:25 vm01.local ceph-mon[48597]: pgmap v3318: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:27 vm08.local ceph-mon[50506]: pgmap v3319: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:27 vm01.local ceph-mon[48597]: pgmap v3319: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:29 vm08.local ceph-mon[50506]: pgmap v3320: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:11:29.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:29 vm01.local ceph-mon[48597]: pgmap v3320: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:11:29.634 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:11:29.635 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:11:29.662 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:11:29.662 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:11:31.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:31 vm08.local ceph-mon[50506]: pgmap v3321: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:31.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:31 vm01.local ceph-mon[48597]: pgmap v3321: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:33 vm08.local ceph-mon[50506]: pgmap v3322: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:11:33.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:33 vm01.local ceph-mon[48597]: pgmap v3322: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:11:34.664 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:11:34.664 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:11:34.692 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:11:34.692 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:11:35.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:35 vm08.local ceph-mon[50506]: pgmap v3323: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:35.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:35 vm01.local ceph-mon[48597]: pgmap v3323: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:37 vm08.local ceph-mon[50506]: pgmap v3324: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:37 vm01.local ceph-mon[48597]: pgmap v3324: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:39 vm08.local ceph-mon[50506]: pgmap v3325: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:11:39.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:39 vm01.local ceph-mon[48597]: pgmap v3325: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:11:39.694 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:11:39.695 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:11:39.722 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:11:39.722 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:11:41.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:41 vm08.local ceph-mon[50506]: pgmap v3326: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:41.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:41 vm01.local ceph-mon[48597]: pgmap v3326: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:42.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:42 vm08.local ceph-mon[50506]: pgmap v3327: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:11:42.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:42 vm01.local ceph-mon[48597]: pgmap v3327: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:11:44.724 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:11:44.724 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:11:44.752 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:11:44.752 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:11:44.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:44 vm08.local ceph-mon[50506]: pgmap v3328: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:44.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:44 vm01.local ceph-mon[48597]: pgmap v3328: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:46.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:46 vm08.local ceph-mon[50506]: pgmap v3329: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:46.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:46 vm01.local ceph-mon[48597]: pgmap v3329: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:48.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:48 vm08.local ceph-mon[50506]: pgmap v3330: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:11:48.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:48 vm01.local ceph-mon[48597]: pgmap v3330: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:11:49.754 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:11:49.754 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:11:49.785 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:11:49.786 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:11:50.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:50 vm01.local ceph-mon[48597]: pgmap v3331: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:51.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:50 vm08.local ceph-mon[50506]: pgmap v3331: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:53.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:52 vm08.local ceph-mon[50506]: pgmap v3332: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:11:53.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:52 vm01.local ceph-mon[48597]: pgmap v3332: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:11:54.787 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:11:54.788 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:11:54.815 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:11:54.815 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:11:55.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:54 vm08.local ceph-mon[50506]: pgmap v3333: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:55.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:54 vm01.local ceph-mon[48597]: pgmap v3333: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:56.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:56 vm08.local ceph-mon[50506]: pgmap v3334: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:57.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:56 vm01.local ceph-mon[48597]: pgmap v3334: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:11:59.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:58 vm08.local ceph-mon[50506]: pgmap v3335: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:11:59.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:58 vm01.local ceph-mon[48597]: pgmap v3335: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:11:59.816 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:11:59.817 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:11:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:59 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:11:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:59 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:11:59.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:11:59 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:11:59.844 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:11:59.845 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:11:59.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:59 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:11:59.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:59 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:11:59.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:11:59 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:12:00.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:00 vm01.local ceph-mon[48597]: pgmap v3336: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:00.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:00 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:12:00.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:00 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:12:01.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:00 vm08.local ceph-mon[50506]: pgmap v3336: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:01.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:00 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:12:01.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:00 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:12:03.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:02 vm08.local ceph-mon[50506]: pgmap v3337: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:12:03.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:02 vm01.local ceph-mon[48597]: pgmap v3337: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:12:04.847 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:12:04.847 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:12:04.874 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:12:04.874 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:12:05.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:04 vm08.local ceph-mon[50506]: pgmap v3338: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:05.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:04 vm01.local ceph-mon[48597]: pgmap v3338: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:07.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:06 vm08.local ceph-mon[50506]: pgmap v3339: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:07.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:06 vm01.local ceph-mon[48597]: pgmap v3339: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:09.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:08 vm08.local ceph-mon[50506]: pgmap v3340: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:12:09.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:08 vm01.local ceph-mon[48597]: pgmap v3340: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:12:09.876 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:12:09.877 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:12:09.906 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:12:09.906 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:12:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:10 vm08.local ceph-mon[50506]: pgmap v3341: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:12:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:12:11.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:10 vm01.local ceph-mon[48597]: pgmap v3341: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:11.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:12:11.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:12:13.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:12 vm08.local ceph-mon[50506]: pgmap v3342: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:12:13.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:12 vm01.local ceph-mon[48597]: pgmap v3342: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:12:14.908 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:12:14.908 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:12:14.937 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:12:14.938 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:12:15.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:14 vm08.local ceph-mon[50506]: pgmap v3343: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:15.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:14 vm01.local ceph-mon[48597]: pgmap v3343: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:17.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:16 vm08.local ceph-mon[50506]: pgmap v3344: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:17.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:16 vm01.local ceph-mon[48597]: pgmap v3344: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:19.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:18 vm08.local ceph-mon[50506]: pgmap v3345: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:12:19.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:18 vm01.local ceph-mon[48597]: pgmap v3345: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:12:19.939 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:12:19.940 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:12:19.966 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:12:19.966 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:12:20.888 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:20 vm01.local ceph-mon[48597]: pgmap v3346: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:21.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:20 vm08.local ceph-mon[50506]: pgmap v3346: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:23.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:22 vm08.local ceph-mon[50506]: pgmap v3347: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:12:23.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:22 vm01.local ceph-mon[48597]: pgmap v3347: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:12:24.968 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:12:24.968 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:12:24.995 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:12:24.995 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:12:25.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:24 vm08.local ceph-mon[50506]: pgmap v3348: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:25.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:24 vm01.local ceph-mon[48597]: pgmap v3348: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:27.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:26 vm08.local ceph-mon[50506]: pgmap v3349: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:27.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:26 vm01.local ceph-mon[48597]: pgmap v3349: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:29.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:28 vm08.local ceph-mon[50506]: pgmap v3350: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:12:29.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:28 vm01.local ceph-mon[48597]: pgmap v3350: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:12:29.997 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:12:29.998 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:12:30.170 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:12:30.170 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:12:31.335 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:30 vm01.local ceph-mon[48597]: pgmap v3351: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:31.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:30 vm08.local ceph-mon[50506]: pgmap v3351: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:33.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:32 vm08.local ceph-mon[50506]: pgmap v3352: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:12:33.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:32 vm01.local ceph-mon[48597]: pgmap v3352: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:12:35.172 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:12:35.172 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:12:35.198 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:12:35.198 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:12:35.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:34 vm08.local ceph-mon[50506]: pgmap v3353: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:35.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:34 vm01.local ceph-mon[48597]: pgmap v3353: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:37.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:36 vm08.local ceph-mon[50506]: pgmap v3354: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:37.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:36 vm01.local ceph-mon[48597]: pgmap v3354: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:39.190 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:38 vm01.local ceph-mon[48597]: pgmap v3355: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:12:39.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:38 vm08.local ceph-mon[50506]: pgmap v3355: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:12:40.199 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:12:40.200 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:12:40.228 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:12:40.228 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:12:41.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:40 vm08.local ceph-mon[50506]: pgmap v3356: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:40 vm01.local ceph-mon[48597]: pgmap v3356: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:43.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:42 vm08.local ceph-mon[50506]: pgmap v3357: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:12:43.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:42 vm01.local ceph-mon[48597]: pgmap v3357: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:12:45.229 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:12:45.230 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:12:45.257 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:12:45.257 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:12:45.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:44 vm08.local ceph-mon[50506]: pgmap v3358: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:45.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:44 vm01.local ceph-mon[48597]: pgmap v3358: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:47.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:46 vm08.local ceph-mon[50506]: pgmap v3359: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:47.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:46 vm01.local ceph-mon[48597]: pgmap v3359: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:48 vm08.local ceph-mon[50506]: pgmap v3360: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:12:49.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:48 vm01.local ceph-mon[48597]: pgmap v3360: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:12:50.259 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:12:50.259 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:12:50.292 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:12:50.293 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:12:51.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:50 vm08.local ceph-mon[50506]: pgmap v3361: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:50 vm01.local ceph-mon[48597]: pgmap v3361: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:53.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:52 vm08.local ceph-mon[50506]: pgmap v3362: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:12:53.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:52 vm01.local ceph-mon[48597]: pgmap v3362: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:12:55.295 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:12:55.295 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:12:55.326 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:12:55.327 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:12:55.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:54 vm08.local ceph-mon[50506]: pgmap v3363: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:54 vm01.local ceph-mon[48597]: pgmap v3363: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:57.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:57 vm08.local ceph-mon[50506]: pgmap v3364: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:57.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:56 vm01.local ceph-mon[48597]: pgmap v3364: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:12:59.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:12:59 vm08.local ceph-mon[50506]: pgmap v3365: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:12:59.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:12:59 vm01.local ceph-mon[48597]: pgmap v3365: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:13:00.011 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:00 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:13:00.011 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:00 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:13:00.011 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:00 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:13:00.011 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:00 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:13:00.011 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:00 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:13:00.011 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:00 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:13:00.011 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:00 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:13:00.011 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:00 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:13:00.328 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:13:00.329 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:13:00.355 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:13:00.356 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:13:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:01 vm01.local ceph-mon[48597]: pgmap v3366: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:01 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:13:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:01 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:13:01.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:01 vm08.local ceph-mon[50506]: pgmap v3366: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:01.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:01 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:13:01.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:01 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:13:03.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:03 vm08.local ceph-mon[50506]: pgmap v3367: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:13:03.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:03 vm01.local ceph-mon[48597]: pgmap v3367: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:13:05.357 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:13:05.358 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:13:05.384 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:13:05.384 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:13:05.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:05 vm08.local ceph-mon[50506]: pgmap v3368: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:05.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:05 vm01.local ceph-mon[48597]: pgmap v3368: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:07 vm08.local ceph-mon[50506]: pgmap v3369: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:07 vm01.local ceph-mon[48597]: pgmap v3369: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:09 vm08.local ceph-mon[50506]: pgmap v3370: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:13:09.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:09 vm01.local ceph-mon[48597]: pgmap v3370: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:13:10.386 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:13:10.386 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:13:10.412 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:13:10.412 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:13:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:11 vm01.local ceph-mon[48597]: pgmap v3371: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:13:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:13:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:11 vm08.local ceph-mon[50506]: pgmap v3371: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:13:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:13:13.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:13 vm08.local ceph-mon[50506]: pgmap v3372: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:13:13.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:13 vm01.local ceph-mon[48597]: pgmap v3372: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:13:15.414 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:13:15.414 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:13:15.442 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:13:15.443 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:13:15.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:15 vm08.local ceph-mon[50506]: pgmap v3373: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:15.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:15 vm01.local ceph-mon[48597]: pgmap v3373: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:17 vm08.local ceph-mon[50506]: pgmap v3374: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:17 vm01.local ceph-mon[48597]: pgmap v3374: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:19.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:19 vm08.local ceph-mon[50506]: pgmap v3375: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:13:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:19 vm01.local ceph-mon[48597]: pgmap v3375: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:13:20.444 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:13:20.444 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:13:20.472 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:13:20.473 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:13:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:21 vm01.local ceph-mon[48597]: pgmap v3376: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:21.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:21 vm08.local ceph-mon[50506]: pgmap v3376: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:23 vm08.local ceph-mon[50506]: pgmap v3377: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:13:23.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:23 vm01.local ceph-mon[48597]: pgmap v3377: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:13:25.474 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:13:25.475 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:13:25.691 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:13:25.691 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:13:25.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:25 vm08.local ceph-mon[50506]: pgmap v3378: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:25.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:25 vm01.local ceph-mon[48597]: pgmap v3378: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:26.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:26 vm08.local ceph-mon[50506]: pgmap v3379: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:27.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:26 vm01.local ceph-mon[48597]: pgmap v3379: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:29.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:28 vm08.local ceph-mon[50506]: pgmap v3380: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:29.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:28 vm01.local ceph-mon[48597]: pgmap v3380: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:30.693 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:13:30.693 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:13:30.718 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:13:30.719 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:13:31.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:30 vm08.local ceph-mon[50506]: pgmap v3381: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T12:13:31.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:30 vm01.local ceph-mon[48597]: pgmap v3381: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T12:13:33.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:32 vm08.local ceph-mon[50506]: pgmap v3382: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:33.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:32 vm01.local ceph-mon[48597]: pgmap v3382: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:35.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:34 vm08.local ceph-mon[50506]: pgmap v3383: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T12:13:35.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:34 vm01.local ceph-mon[48597]: pgmap v3383: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T12:13:35.720 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:13:35.721 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:13:35.748 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:13:35.749 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:13:37.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:36 vm08.local ceph-mon[50506]: pgmap v3384: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T12:13:37.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:36 vm01.local ceph-mon[48597]: pgmap v3384: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-10T12:13:39.193 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:38 vm01.local ceph-mon[48597]: pgmap v3385: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:39.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:38 vm08.local ceph-mon[50506]: pgmap v3385: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:40.750 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:13:40.751 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:13:40.777 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:13:40.777 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:13:41.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:40 vm08.local ceph-mon[50506]: pgmap v3386: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:40 vm01.local ceph-mon[48597]: pgmap v3386: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:43.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:42 vm08.local ceph-mon[50506]: pgmap v3387: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:13:43.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:42 vm01.local ceph-mon[48597]: pgmap v3387: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:13:45.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:44 vm08.local ceph-mon[50506]: pgmap v3388: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:45.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:44 vm01.local ceph-mon[48597]: pgmap v3388: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:45.778 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:13:45.779 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:13:45.805 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:13:45.805 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:13:47.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:46 vm08.local ceph-mon[50506]: pgmap v3389: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:47.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:46 vm01.local ceph-mon[48597]: pgmap v3389: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:48 vm08.local ceph-mon[50506]: pgmap v3390: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:13:49.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:48 vm01.local ceph-mon[48597]: pgmap v3390: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:13:50.806 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:13:50.807 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:13:50.833 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:13:50.833 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:13:51.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:50 vm08.local ceph-mon[50506]: pgmap v3391: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:50 vm01.local ceph-mon[48597]: pgmap v3391: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:53.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:52 vm08.local ceph-mon[50506]: pgmap v3392: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:13:53.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:52 vm01.local ceph-mon[48597]: pgmap v3392: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:13:55.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:54 vm08.local ceph-mon[50506]: pgmap v3393: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:54 vm01.local ceph-mon[48597]: pgmap v3393: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:55.835 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:13:55.835 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:13:55.860 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:13:55.861 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:13:57.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:56 vm08.local ceph-mon[50506]: pgmap v3394: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:57.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:56 vm01.local ceph-mon[48597]: pgmap v3394: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:13:59.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:13:58 vm08.local ceph-mon[50506]: pgmap v3395: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:13:59.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:13:58 vm01.local ceph-mon[48597]: pgmap v3395: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:14:00.862 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:14:00.863 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:14:00.888 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:14:00.889 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:14:01.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:00 vm08.local ceph-mon[50506]: pgmap v3396: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:01.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:00 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:14:01.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:00 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:14:01.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:00 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:14:01.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:00 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:14:01.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:00 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:14:01.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:00 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:14:01.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:00 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:14:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:00 vm01.local ceph-mon[48597]: pgmap v3396: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:00 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:14:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:00 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:14:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:00 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:14:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:00 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:14:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:00 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:14:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:00 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:14:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:00 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:14:03.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:03 vm08.local ceph-mon[50506]: pgmap v3397: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:14:03.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:03 vm01.local ceph-mon[48597]: pgmap v3397: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:14:05.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:05 vm08.local ceph-mon[50506]: pgmap v3398: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:05.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:05 vm01.local ceph-mon[48597]: pgmap v3398: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:05.890 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:14:05.891 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:14:06.099 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:14:06.100 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:14:07.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:07 vm08.local ceph-mon[50506]: pgmap v3399: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:07.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:07 vm01.local ceph-mon[48597]: pgmap v3399: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:09.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:09 vm08.local ceph-mon[50506]: pgmap v3400: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:14:09.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:09 vm01.local ceph-mon[48597]: pgmap v3400: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:14:11.101 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:14:11.102 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:14:11.128 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:14:11.129 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:14:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:11 vm01.local ceph-mon[48597]: pgmap v3401: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:14:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:14:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:11 vm08.local ceph-mon[50506]: pgmap v3401: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:14:11.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:14:13.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:13 vm08.local ceph-mon[50506]: pgmap v3402: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:14:13.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:13 vm01.local ceph-mon[48597]: pgmap v3402: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:14:15.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:15 vm08.local ceph-mon[50506]: pgmap v3403: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:15.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:15 vm01.local ceph-mon[48597]: pgmap v3403: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:16.130 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:14:16.131 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:14:16.158 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:14:16.158 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:14:17.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:17 vm08.local ceph-mon[50506]: pgmap v3404: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:17.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:17 vm01.local ceph-mon[48597]: pgmap v3404: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:19.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:19 vm08.local ceph-mon[50506]: pgmap v3405: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:14:19.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:19 vm01.local ceph-mon[48597]: pgmap v3405: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:14:21.160 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:14:21.160 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:14:21.187 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:14:21.188 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:14:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:21 vm01.local ceph-mon[48597]: pgmap v3406: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:21.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:21 vm08.local ceph-mon[50506]: pgmap v3406: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:23 vm08.local ceph-mon[50506]: pgmap v3407: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:14:23.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:23 vm01.local ceph-mon[48597]: pgmap v3407: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:14:25.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:25 vm08.local ceph-mon[50506]: pgmap v3408: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:25.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:25 vm01.local ceph-mon[48597]: pgmap v3408: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:26.189 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:14:26.190 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:14:26.217 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:14:26.218 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:14:27.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:27 vm08.local ceph-mon[50506]: pgmap v3409: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:27 vm01.local ceph-mon[48597]: pgmap v3409: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:29 vm08.local ceph-mon[50506]: pgmap v3410: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:14:29.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:29 vm01.local ceph-mon[48597]: pgmap v3410: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:14:31.219 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:14:31.220 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:14:31.249 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:14:31.250 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:14:31.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:31 vm08.local ceph-mon[50506]: pgmap v3411: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:31.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:31 vm01.local ceph-mon[48597]: pgmap v3411: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:33 vm08.local ceph-mon[50506]: pgmap v3412: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:14:33.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:33 vm01.local ceph-mon[48597]: pgmap v3412: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:14:35.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:34 vm08.local ceph-mon[50506]: pgmap v3413: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:35.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:34 vm01.local ceph-mon[48597]: pgmap v3413: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:36.251 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:14:36.251 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:14:36.278 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:14:36.278 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:14:37.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:36 vm08.local ceph-mon[50506]: pgmap v3414: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:37.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:36 vm01.local ceph-mon[48597]: pgmap v3414: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:39.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:38 vm08.local ceph-mon[50506]: pgmap v3415: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:14:39.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:38 vm01.local ceph-mon[48597]: pgmap v3415: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:14:41.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:40 vm08.local ceph-mon[50506]: pgmap v3416: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:41.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:40 vm01.local ceph-mon[48597]: pgmap v3416: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:41.279 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:14:41.280 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:14:41.307 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:14:41.308 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:14:43.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:42 vm08.local ceph-mon[50506]: pgmap v3417: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:14:43.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:42 vm01.local ceph-mon[48597]: pgmap v3417: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:14:45.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:44 vm08.local ceph-mon[50506]: pgmap v3418: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:45.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:44 vm01.local ceph-mon[48597]: pgmap v3418: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:46.309 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:14:46.310 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:14:46.335 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:14:46.336 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:14:47.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:46 vm08.local ceph-mon[50506]: pgmap v3419: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:47.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:46 vm01.local ceph-mon[48597]: pgmap v3419: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:49.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:48 vm08.local ceph-mon[50506]: pgmap v3420: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:14:49.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:48 vm01.local ceph-mon[48597]: pgmap v3420: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:14:51.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:50 vm08.local ceph-mon[50506]: pgmap v3421: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:51.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:50 vm01.local ceph-mon[48597]: pgmap v3421: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:51.337 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:14:51.338 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:14:51.364 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:14:51.364 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:14:53.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:52 vm08.local ceph-mon[50506]: pgmap v3422: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:14:53.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:52 vm01.local ceph-mon[48597]: pgmap v3422: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:14:55.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:54 vm01.local ceph-mon[48597]: pgmap v3423: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:55.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:54 vm08.local ceph-mon[50506]: pgmap v3423: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:56.366 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:14:56.366 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:14:56.393 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:14:56.394 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:14:57.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:56 vm01.local ceph-mon[48597]: pgmap v3424: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:57.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:56 vm08.local ceph-mon[50506]: pgmap v3424: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:14:59.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:14:58 vm01.local ceph-mon[48597]: pgmap v3425: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:14:59.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:14:58 vm08.local ceph-mon[50506]: pgmap v3425: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:15:00.901 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:00 vm01.local ceph-mon[48597]: pgmap v3426: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:00.901 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:00 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:15:00.901 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:00 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:15:00.901 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:00 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:15:00.914 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:00 vm08.local ceph-mon[50506]: pgmap v3426: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:00.915 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:00 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:15:00.915 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:00 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:15:00.915 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:00 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:15:01.396 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:15:01.400 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:15:01.449 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:15:01.449 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:15:02.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:02 vm08.local ceph-mon[50506]: pgmap v3427: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:15:02.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:02 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:15:02.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:02 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:15:02.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:02 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:15:02.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:02 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:15:02.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:02 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:15:02.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:02 vm01.local ceph-mon[48597]: pgmap v3427: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:15:02.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:02 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:15:02.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:02 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:15:02.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:02 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:15:02.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:02 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:15:02.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:02 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:15:04.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:04 vm08.local ceph-mon[50506]: pgmap v3428: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:04.868 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:04 vm01.local ceph-mon[48597]: pgmap v3428: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:06.451 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:15:06.451 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:15:06.479 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:15:06.480 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:15:06.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:06 vm08.local ceph-mon[50506]: pgmap v3429: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:07.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:06 vm01.local ceph-mon[48597]: pgmap v3429: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:09.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:08 vm08.local ceph-mon[50506]: pgmap v3430: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:15:09.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:08 vm01.local ceph-mon[48597]: pgmap v3430: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:15:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:10 vm01.local ceph-mon[48597]: pgmap v3431: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:15:10.993 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:15:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:10 vm08.local ceph-mon[50506]: pgmap v3431: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:15:11.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:15:11.481 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:15:11.481 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:15:11.507 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:15:11.508 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:15:13.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:12 vm08.local ceph-mon[50506]: pgmap v3432: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:15:13.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:12 vm01.local ceph-mon[48597]: pgmap v3432: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:15:15.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:14 vm08.local ceph-mon[50506]: pgmap v3433: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:15.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:14 vm01.local ceph-mon[48597]: pgmap v3433: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:16.509 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:15:16.509 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:15:16.536 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:15:16.537 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:15:16.844 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:16 vm08.local ceph-mon[50506]: pgmap v3434: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:17.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:16 vm01.local ceph-mon[48597]: pgmap v3434: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:19.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:18 vm08.local ceph-mon[50506]: pgmap v3435: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:15:19.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:18 vm01.local ceph-mon[48597]: pgmap v3435: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:15:20.906 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:20 vm01.local ceph-mon[48597]: pgmap v3436: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:21.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:20 vm08.local ceph-mon[50506]: pgmap v3436: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:21.539 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:15:21.540 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:15:21.577 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:15:21.577 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:15:23.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:22 vm08.local ceph-mon[50506]: pgmap v3437: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:15:23.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:22 vm01.local ceph-mon[48597]: pgmap v3437: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:15:25.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:24 vm08.local ceph-mon[50506]: pgmap v3438: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:25.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:24 vm01.local ceph-mon[48597]: pgmap v3438: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:26.578 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:15:26.579 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:15:26.615 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:15:26.616 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:15:27.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:26 vm08.local ceph-mon[50506]: pgmap v3439: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:27.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:26 vm01.local ceph-mon[48597]: pgmap v3439: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:29.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:28 vm08.local ceph-mon[50506]: pgmap v3440: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:15:29.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:28 vm01.local ceph-mon[48597]: pgmap v3440: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:15:31.051 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:30 vm01.local ceph-mon[48597]: pgmap v3441: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:31.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:30 vm08.local ceph-mon[50506]: pgmap v3441: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:31.618 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:15:31.618 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:15:31.649 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:15:31.650 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:15:33.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:32 vm08.local ceph-mon[50506]: pgmap v3442: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:15:33.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:32 vm01.local ceph-mon[48597]: pgmap v3442: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:15:35.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:34 vm08.local ceph-mon[50506]: pgmap v3443: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:35.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:34 vm01.local ceph-mon[48597]: pgmap v3443: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:36.651 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:15:36.652 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:15:36.679 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:15:36.680 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:15:37.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:36 vm08.local ceph-mon[50506]: pgmap v3444: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:37.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:36 vm01.local ceph-mon[48597]: pgmap v3444: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:39.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:38 vm08.local ceph-mon[50506]: pgmap v3445: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:15:39.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:38 vm01.local ceph-mon[48597]: pgmap v3445: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:15:41.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:40 vm08.local ceph-mon[50506]: pgmap v3446: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:41.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:40 vm01.local ceph-mon[48597]: pgmap v3446: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:41.681 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:15:41.682 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:15:41.709 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:15:41.709 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:15:43.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:42 vm08.local ceph-mon[50506]: pgmap v3447: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:15:43.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:42 vm01.local ceph-mon[48597]: pgmap v3447: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:15:45.094 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:44 vm08.local ceph-mon[50506]: pgmap v3448: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:45.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:44 vm01.local ceph-mon[48597]: pgmap v3448: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:46.711 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:15:46.711 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:15:46.739 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:15:46.740 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:15:47.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:46 vm01.local ceph-mon[48597]: pgmap v3449: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:47.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:46 vm08.local ceph-mon[50506]: pgmap v3449: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:49.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:48 vm01.local ceph-mon[48597]: pgmap v3450: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:15:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:48 vm08.local ceph-mon[50506]: pgmap v3450: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:15:51.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:50 vm01.local ceph-mon[48597]: pgmap v3451: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:51.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:50 vm08.local ceph-mon[50506]: pgmap v3451: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:51.741 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:15:51.741 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:15:51.767 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:15:51.767 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:15:53.118 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:52 vm01.local ceph-mon[48597]: pgmap v3452: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:15:53.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:52 vm08.local ceph-mon[50506]: pgmap v3452: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:15:55.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:54 vm08.local ceph-mon[50506]: pgmap v3453: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:54 vm01.local ceph-mon[48597]: pgmap v3453: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:56.769 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:15:56.769 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:15:56.796 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:15:56.796 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:15:57.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:56 vm08.local ceph-mon[50506]: pgmap v3454: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:57.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:56 vm01.local ceph-mon[48597]: pgmap v3454: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:15:59.202 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:15:58 vm01.local ceph-mon[48597]: pgmap v3455: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:15:59.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:15:58 vm08.local ceph-mon[50506]: pgmap v3455: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:16:01.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:00 vm08.local ceph-mon[50506]: pgmap v3456: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:00 vm01.local ceph-mon[48597]: pgmap v3456: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:01.798 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:16:01.798 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:16:01.825 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:16:01.825 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:16:03.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:02 vm08.local ceph-mon[50506]: pgmap v3457: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:16:03.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:02 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:16:03.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:02 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:16:03.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:02 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:16:03.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:02 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:16:03.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:02 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:16:03.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:02 vm01.local ceph-mon[48597]: pgmap v3457: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:16:03.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:02 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:16:03.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:02 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:16:03.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:02 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:16:03.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:02 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:16:03.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:02 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:16:05.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:04 vm08.local ceph-mon[50506]: pgmap v3458: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:05.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:04 vm01.local ceph-mon[48597]: pgmap v3458: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:06.827 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:16:06.827 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:16:06.855 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:16:06.856 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:16:07.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:06 vm08.local ceph-mon[50506]: pgmap v3459: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:07.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:06 vm01.local ceph-mon[48597]: pgmap v3459: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:09.203 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:08 vm01.local ceph-mon[48597]: pgmap v3460: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:16:09.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:08 vm08.local ceph-mon[50506]: pgmap v3460: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:16:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:10 vm08.local ceph-mon[50506]: pgmap v3461: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:16:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:10 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:16:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:10 vm01.local ceph-mon[48597]: pgmap v3461: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:16:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:10 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:16:11.857 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:16:11.858 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:16:11.884 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:16:11.884 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:16:13.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:12 vm08.local ceph-mon[50506]: pgmap v3462: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:16:13.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:12 vm01.local ceph-mon[48597]: pgmap v3462: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:16:15.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:14 vm08.local ceph-mon[50506]: pgmap v3463: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:15.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:14 vm01.local ceph-mon[48597]: pgmap v3463: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:16.886 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:16:16.886 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:16:16.912 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:16:16.913 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:16:17.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:16 vm08.local ceph-mon[50506]: pgmap v3464: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:17.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:16 vm01.local ceph-mon[48597]: pgmap v3464: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:19.203 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:18 vm01.local ceph-mon[48597]: pgmap v3465: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:16:19.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:18 vm08.local ceph-mon[50506]: pgmap v3465: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:16:21.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:20 vm08.local ceph-mon[50506]: pgmap v3466: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:20 vm01.local ceph-mon[48597]: pgmap v3466: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:21.914 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:16:21.915 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:16:21.941 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:16:21.942 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:16:23.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:22 vm08.local ceph-mon[50506]: pgmap v3467: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:16:23.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:22 vm01.local ceph-mon[48597]: pgmap v3467: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:16:25.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:24 vm08.local ceph-mon[50506]: pgmap v3468: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:25.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:24 vm01.local ceph-mon[48597]: pgmap v3468: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:26.943 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:16:26.944 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:16:26.970 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:16:26.970 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:16:27.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:26 vm08.local ceph-mon[50506]: pgmap v3469: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:27.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:26 vm01.local ceph-mon[48597]: pgmap v3469: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:29.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:28 vm08.local ceph-mon[50506]: pgmap v3470: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:16:29.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:28 vm01.local ceph-mon[48597]: pgmap v3470: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:16:31.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:30 vm08.local ceph-mon[50506]: pgmap v3471: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:31.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:30 vm01.local ceph-mon[48597]: pgmap v3471: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:31.971 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:16:31.972 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:16:31.999 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:16:32.000 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:16:33.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:32 vm08.local ceph-mon[50506]: pgmap v3472: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:16:33.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:32 vm01.local ceph-mon[48597]: pgmap v3472: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:16:35.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:34 vm08.local ceph-mon[50506]: pgmap v3473: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:35.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:34 vm01.local ceph-mon[48597]: pgmap v3473: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:37.001 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:16:37.002 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:16:37.027 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:16:37.028 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:16:37.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:36 vm08.local ceph-mon[50506]: pgmap v3474: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:37.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:36 vm01.local ceph-mon[48597]: pgmap v3474: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:39.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:38 vm08.local ceph-mon[50506]: pgmap v3475: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:16:39.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:38 vm01.local ceph-mon[48597]: pgmap v3475: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:16:41.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:40 vm08.local ceph-mon[50506]: pgmap v3476: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:41.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:40 vm01.local ceph-mon[48597]: pgmap v3476: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:42.029 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:16:42.030 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:16:42.055 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:16:42.056 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:16:43.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:43 vm08.local ceph-mon[50506]: pgmap v3477: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:16:43.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:43 vm01.local ceph-mon[48597]: pgmap v3477: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:16:45.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:45 vm08.local ceph-mon[50506]: pgmap v3478: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:45.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:45 vm01.local ceph-mon[48597]: pgmap v3478: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:47.057 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:16:47.058 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:16:47.084 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:16:47.085 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:16:47.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:47 vm08.local ceph-mon[50506]: pgmap v3479: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:47.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:47 vm01.local ceph-mon[48597]: pgmap v3479: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:49.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:49 vm08.local ceph-mon[50506]: pgmap v3480: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:16:49.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:49 vm01.local ceph-mon[48597]: pgmap v3480: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:16:51.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:51 vm08.local ceph-mon[50506]: pgmap v3481: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:51.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:51 vm01.local ceph-mon[48597]: pgmap v3481: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:52.086 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:16:52.086 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:16:52.112 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:16:52.113 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:16:53.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:53 vm08.local ceph-mon[50506]: pgmap v3482: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:16:53.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:53 vm01.local ceph-mon[48597]: pgmap v3482: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:16:55.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:55 vm08.local ceph-mon[50506]: pgmap v3483: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:55.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:55 vm01.local ceph-mon[48597]: pgmap v3483: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:57.114 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:16:57.115 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:16:57.143 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:16:57.144 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:16:57.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:57 vm08.local ceph-mon[50506]: pgmap v3484: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:57.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:57 vm01.local ceph-mon[48597]: pgmap v3484: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:16:59.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:16:59 vm08.local ceph-mon[50506]: pgmap v3485: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:16:59.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:16:59 vm01.local ceph-mon[48597]: pgmap v3485: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:17:01.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:17:01 vm08.local ceph-mon[50506]: pgmap v3486: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:17:01.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:17:01 vm01.local ceph-mon[48597]: pgmap v3486: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:17:02.145 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:17:02.146 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:17:02.171 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:17:02.172 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:17:03.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:17:03 vm08.local ceph-mon[50506]: pgmap v3487: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:17:03.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:17:03 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:17:03.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:17:03 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:17:03.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:17:03 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:17:03.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:17:03 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:17:03.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:17:03 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:17:03.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:17:03 vm01.local ceph-mon[48597]: pgmap v3487: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:17:03.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:17:03 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:17:03.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:17:03 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:17:03.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:17:03 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:17:03.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:17:03 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:17:03.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:17:03 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' 2026-03-10T12:17:05.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:17:05 vm08.local ceph-mon[50506]: pgmap v3488: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:17:05.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:17:05 vm01.local ceph-mon[48597]: pgmap v3488: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:17:07.173 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:17:07.174 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:17:07.199 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:17:07.200 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:17:07.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:17:07 vm08.local ceph-mon[50506]: pgmap v3489: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:17:07.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:17:07 vm01.local ceph-mon[48597]: pgmap v3489: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:17:09.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:17:09 vm08.local ceph-mon[50506]: pgmap v3490: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:17:09.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:17:09 vm01.local ceph-mon[48597]: pgmap v3490: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:17:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:17:11 vm08.local ceph-mon[50506]: pgmap v3491: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:17:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:17:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:17:11.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:17:11 vm08.local ceph-mon[50506]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:17:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:17:11 vm01.local ceph-mon[48597]: pgmap v3491: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:17:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:17:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:17:11.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:17:11 vm01.local ceph-mon[48597]: from='mgr.14216 192.168.123.101:0/672072430' entity='mgr.vm01.nkkzmc' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.nkkzmc/trash_purge_schedule"}]: dispatch 2026-03-10T12:17:12.201 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:17:12.201 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:17:12.227 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:17:12.227 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:17:13.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:17:13 vm08.local ceph-mon[50506]: pgmap v3492: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:17:13.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:17:13 vm01.local ceph-mon[48597]: pgmap v3492: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:17:15.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:17:15 vm08.local ceph-mon[50506]: pgmap v3493: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:17:15.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:17:15 vm01.local ceph-mon[48597]: pgmap v3493: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:17:17.229 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:17:17.229 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:17:17.256 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:17:17.257 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:17:17.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:17:17 vm01.local ceph-mon[48597]: pgmap v3494: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:17:17.502 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:17:17 vm08.local ceph-mon[50506]: pgmap v3494: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:17:19.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:17:19 vm01.local ceph-mon[48597]: pgmap v3495: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:17:19.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:17:19 vm08.local ceph-mon[50506]: pgmap v3495: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:17:21.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:17:21 vm01.local ceph-mon[48597]: pgmap v3496: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:17:21.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:17:21 vm08.local ceph-mon[50506]: pgmap v3496: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:17:22.258 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:17:22.258 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:17:22.286 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:17:22.287 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:17:23.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:17:23 vm01.local ceph-mon[48597]: pgmap v3497: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:17:23.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:17:23 vm08.local ceph-mon[50506]: pgmap v3497: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:17:25.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:17:25 vm08.local ceph-mon[50506]: pgmap v3498: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:17:25.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:17:25 vm01.local ceph-mon[48597]: pgmap v3498: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:17:27.288 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:17:27.288 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:17:27.316 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:17:27.316 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:17:27.536 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:17:27 vm08.local ceph-mon[50506]: pgmap v3499: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:17:27.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:17:27 vm01.local ceph-mon[48597]: pgmap v3499: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:17:29.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:17:29 vm08.local ceph-mon[50506]: pgmap v3500: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:17:29.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:17:29 vm01.local ceph-mon[48597]: pgmap v3500: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:17:31.368 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:17:31 vm01.local ceph-mon[48597]: pgmap v3501: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:17:31.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:17:31 vm08.local ceph-mon[50506]: pgmap v3501: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:17:32.317 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:17:32.318 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:17:32.344 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:17:32.344 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:17:33.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:17:33 vm08.local ceph-mon[50506]: pgmap v3502: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:17:33.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:17:33 vm01.local ceph-mon[48597]: pgmap v3502: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:17:35.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:17:35 vm08.local ceph-mon[50506]: pgmap v3503: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:17:35.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:17:35 vm01.local ceph-mon[48597]: pgmap v3503: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:17:37.345 INFO:teuthology.orchestra.run.vm01.stderr:++ hostname 2026-03-10T12:17:37.346 INFO:teuthology.orchestra.run.vm01.stderr:+ mount -t nfs vm01.local:/fake /mnt/foo -o sync 2026-03-10T12:17:37.373 INFO:teuthology.orchestra.run.vm01.stderr:mount.nfs: mount system call failed 2026-03-10T12:17:37.373 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-10T12:17:37.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:17:37 vm08.local ceph-mon[50506]: pgmap v3504: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:17:37.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:17:37 vm01.local ceph-mon[48597]: pgmap v3504: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-10T12:17:39.594 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 10 12:17:39 vm08.local ceph-mon[50506]: pgmap v3505: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:17:39.618 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 12:17:39 vm01.local ceph-mon[48597]: pgmap v3505: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-10T12:17:40.272 DEBUG:teuthology.exit:Got signal 15; running 1 handler... 2026-03-10T12:17:40.272 DEBUG:teuthology.exit:Finished running handlers