2026-03-10T05:48:49.501 INFO:root:teuthology version: 1.2.4.dev6+g1c580df7a 2026-03-10T05:48:49.506 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-10T05:48:49.532 INFO:teuthology.run:Config: archive_path: /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/921 branch: squid description: orch/cephadm/smoke-roleless/{0-distro/centos_9.stream_runc 1-start 2-services/client-keyring 3-final} email: null first_in_suite: false flavor: default job_id: '921' last_in_suite: false machine_type: vps name: kyr-2026-03-10_01:00:38-orch-squid-none-default-vps no_nested_subset: false openstack: - volumes: count: 4 size: 10 os_type: centos os_version: 9.stream overrides: admin_socket: branch: squid ansible.cephlab: branch: main skip_tags: nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs vars: timezone: UTC ceph: conf: mgr: debug mgr: 20 debug ms: 1 mon: debug mon: 20 debug ms: 1 debug paxos: 20 osd: debug ms: 1 debug osd: 20 osd mclock iops capacity threshold hdd: 49000 osd shutdown pgref assert: true flavor: default log-ignorelist: - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - CEPHADM_DAEMON_PLACE_FAIL - CEPHADM_FAILED_DAEMON log-only-match: - CEPHADM_ sha1: e911bdebe5c8faa3800735d1568fcdca65db60df ceph-deploy: conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: {} install: ceph: flavor: default sha1: e911bdebe5c8faa3800735d1568fcdca65db60df extra_system_packages: deb: - python3-xmltodict - python3-jmespath rpm: - bzip2 - perl-Test-Harness - python3-xmltodict - python3-jmespath selinux: allowlist: - scontext=system_u:system_r:logrotate_t:s0 workunit: branch: tt-squid sha1: 75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b owner: kyr priority: 1000 repo: https://github.com/ceph/ceph.git roles: - - host.a - client.0 - - host.b - client.1 seed: 8043 sha1: e911bdebe5c8faa3800735d1568fcdca65db60df sleep_before_teardown: 0 subset: 1/64 suite: orch suite_branch: tt-squid suite_path: /home/teuthos/src/github.com_kshtsk_ceph_75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b/qa suite_relpath: qa suite_repo: https://github.com/kshtsk/ceph.git suite_sha1: 75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b targets: vm01.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBDysC8eZF20UJXhAR/xQ08uDjvTJMaUaBcpCEUGv5qYQuA4lDPCvk5r6vZH6lC3Q+LSRWcnpL9eyQoMNQqRZPv0= vm09.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBKo6nSd8/esxm2uUf2rKVPGxrNn//p0K0pj9FVAFxxQ7JA18myq+9Znha+eDoml7KhfqUGTufgu7EvQAwSNM0sU= tasks: - pexec: all: - sudo dnf remove nvme-cli -y - sudo dnf install runc nvmetcli nvme-cli -y - sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf - sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf - cephadm: roleless: true - cephadm.shell: host.a: - ceph orch status - ceph orch ps - ceph orch ls - ceph orch host ls - ceph orch device ls - cephadm.shell: host.a: - ceph orch host label add `hostname` foo - ceph auth get-or-create client.foo mon 'allow r' - ceph orch client-keyring set client.foo label:foo --mode 770 --owner 11111:22222 - exec: host.a: - while ! test -e /etc/ceph/ceph.client.foo.keyring ; do sleep 1 ; done - ls -al /etc/ceph/ceph.client.foo.keyring | grep rwxrwx--- - ls -al /etc/ceph/ceph.client.foo.keyring | grep 11111 - ls -al /etc/ceph/ceph.client.foo.keyring | grep 22222 - test -e /etc/ceph/ceph.conf - exec: host.b: - test ! -e /etc/ceph/ceph.client.foo.keyring - cephadm.shell: host.b: - ceph orch host label add `hostname` foo - exec: host.b: - while ! test -e /etc/ceph/ceph.client.foo.keyring ; do sleep 1 ; done - ls -al /etc/ceph/ceph.client.foo.keyring | grep rwxrwx--- - ls -al /etc/ceph/ceph.client.foo.keyring | grep 11111 - ls -al /etc/ceph/ceph.client.foo.keyring | grep 22222 - cephadm.shell: host.b: - ceph orch host label rm `hostname` foo - exec: host.b: - while test -e /etc/ceph/ceph.client.foo.keyring ; do sleep 1 ; done - exec: host.a: - test -e /etc/ceph/ceph.client.foo.keyring - cephadm.shell: host.a: - ceph orch client-keyring rm client.foo - exec: host.a: - while test -e /etc/ceph/ceph.client.foo.keyring ; do sleep 1 ; done - cephadm.shell: host.a: - stat -c '%u %g' /var/log/ceph | grep '167 167' - ceph orch status - ceph orch ps - ceph orch ls - ceph orch host ls - ceph orch device ls - ceph orch ls | grep '^osd.all-available-devices ' teuthology: fragments_dropped: [] meta: {} postmerge: [] teuthology_branch: clyso-debian-13 teuthology_repo: https://github.com/clyso/teuthology teuthology_sha1: 1c580df7a9c7c2aadc272da296344fd99f27c444 timestamp: 2026-03-10_01:00:38 tube: vps user: kyr verbose: false worker_log: /home/teuthos/.teuthology/dispatcher/dispatcher.vps.611473 2026-03-10T05:48:49.533 INFO:teuthology.run:suite_path is set to /home/teuthos/src/github.com_kshtsk_ceph_75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b/qa; will attempt to use it 2026-03-10T05:48:49.533 INFO:teuthology.run:Found tasks at /home/teuthos/src/github.com_kshtsk_ceph_75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b/qa/tasks 2026-03-10T05:48:49.533 INFO:teuthology.run_tasks:Running task internal.check_packages... 2026-03-10T05:48:49.534 INFO:teuthology.task.internal:Checking packages... 2026-03-10T05:48:49.534 INFO:teuthology.task.internal:Checking packages for os_type 'centos', flavor 'default' and ceph hash 'e911bdebe5c8faa3800735d1568fcdca65db60df' 2026-03-10T05:48:49.534 WARNING:teuthology.packaging:More than one of ref, tag, branch, or sha1 supplied; using branch 2026-03-10T05:48:49.534 INFO:teuthology.packaging:ref: None 2026-03-10T05:48:49.534 INFO:teuthology.packaging:tag: None 2026-03-10T05:48:49.534 INFO:teuthology.packaging:branch: squid 2026-03-10T05:48:49.534 INFO:teuthology.packaging:sha1: e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T05:48:49.534 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&ref=squid 2026-03-10T05:48:50.284 INFO:teuthology.task.internal:Found packages for ceph version 19.2.3-678.ge911bdeb 2026-03-10T05:48:50.285 INFO:teuthology.run_tasks:Running task internal.buildpackages_prep... 2026-03-10T05:48:50.286 INFO:teuthology.task.internal:no buildpackages task found 2026-03-10T05:48:50.286 INFO:teuthology.run_tasks:Running task internal.save_config... 2026-03-10T05:48:50.286 INFO:teuthology.task.internal:Saving configuration 2026-03-10T05:48:50.291 INFO:teuthology.run_tasks:Running task internal.check_lock... 2026-03-10T05:48:50.292 INFO:teuthology.task.internal.check_lock:Checking locks... 2026-03-10T05:48:50.301 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm01.local', 'description': '/archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/921', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-10 05:47:35.698657', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:01', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBDysC8eZF20UJXhAR/xQ08uDjvTJMaUaBcpCEUGv5qYQuA4lDPCvk5r6vZH6lC3Q+LSRWcnpL9eyQoMNQqRZPv0='} 2026-03-10T05:48:50.310 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm09.local', 'description': '/archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/921', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-10 05:47:35.699291', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:09', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBKo6nSd8/esxm2uUf2rKVPGxrNn//p0K0pj9FVAFxxQ7JA18myq+9Znha+eDoml7KhfqUGTufgu7EvQAwSNM0sU='} 2026-03-10T05:48:50.310 INFO:teuthology.run_tasks:Running task internal.add_remotes... 2026-03-10T05:48:50.311 INFO:teuthology.task.internal:roles: ubuntu@vm01.local - ['host.a', 'client.0'] 2026-03-10T05:48:50.311 INFO:teuthology.task.internal:roles: ubuntu@vm09.local - ['host.b', 'client.1'] 2026-03-10T05:48:50.311 INFO:teuthology.run_tasks:Running task console_log... 2026-03-10T05:48:50.319 DEBUG:teuthology.task.console_log:vm01 does not support IPMI; excluding 2026-03-10T05:48:50.324 DEBUG:teuthology.task.console_log:vm09 does not support IPMI; excluding 2026-03-10T05:48:50.324 DEBUG:teuthology.exit:Installing handler: Handler(exiter=, func=.kill_console_loggers at 0x7f2d0451e170>, signals=[15]) 2026-03-10T05:48:50.324 INFO:teuthology.run_tasks:Running task internal.connect... 2026-03-10T05:48:50.325 INFO:teuthology.task.internal:Opening connections... 2026-03-10T05:48:50.325 DEBUG:teuthology.task.internal:connecting to ubuntu@vm01.local 2026-03-10T05:48:50.326 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm01.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T05:48:50.386 DEBUG:teuthology.task.internal:connecting to ubuntu@vm09.local 2026-03-10T05:48:50.386 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm09.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T05:48:50.446 INFO:teuthology.run_tasks:Running task internal.push_inventory... 2026-03-10T05:48:50.447 DEBUG:teuthology.orchestra.run.vm01:> uname -m 2026-03-10T05:48:50.499 INFO:teuthology.orchestra.run.vm01.stdout:x86_64 2026-03-10T05:48:50.499 DEBUG:teuthology.orchestra.run.vm01:> cat /etc/os-release 2026-03-10T05:48:50.554 INFO:teuthology.orchestra.run.vm01.stdout:NAME="CentOS Stream" 2026-03-10T05:48:50.554 INFO:teuthology.orchestra.run.vm01.stdout:VERSION="9" 2026-03-10T05:48:50.555 INFO:teuthology.orchestra.run.vm01.stdout:ID="centos" 2026-03-10T05:48:50.555 INFO:teuthology.orchestra.run.vm01.stdout:ID_LIKE="rhel fedora" 2026-03-10T05:48:50.555 INFO:teuthology.orchestra.run.vm01.stdout:VERSION_ID="9" 2026-03-10T05:48:50.555 INFO:teuthology.orchestra.run.vm01.stdout:PLATFORM_ID="platform:el9" 2026-03-10T05:48:50.555 INFO:teuthology.orchestra.run.vm01.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-10T05:48:50.555 INFO:teuthology.orchestra.run.vm01.stdout:ANSI_COLOR="0;31" 2026-03-10T05:48:50.555 INFO:teuthology.orchestra.run.vm01.stdout:LOGO="fedora-logo-icon" 2026-03-10T05:48:50.555 INFO:teuthology.orchestra.run.vm01.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-10T05:48:50.555 INFO:teuthology.orchestra.run.vm01.stdout:HOME_URL="https://centos.org/" 2026-03-10T05:48:50.555 INFO:teuthology.orchestra.run.vm01.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-10T05:48:50.555 INFO:teuthology.orchestra.run.vm01.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-10T05:48:50.555 INFO:teuthology.orchestra.run.vm01.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-10T05:48:50.555 INFO:teuthology.lock.ops:Updating vm01.local on lock server 2026-03-10T05:48:50.561 DEBUG:teuthology.orchestra.run.vm09:> uname -m 2026-03-10T05:48:50.578 INFO:teuthology.orchestra.run.vm09.stdout:x86_64 2026-03-10T05:48:50.579 DEBUG:teuthology.orchestra.run.vm09:> cat /etc/os-release 2026-03-10T05:48:50.633 INFO:teuthology.orchestra.run.vm09.stdout:NAME="CentOS Stream" 2026-03-10T05:48:50.633 INFO:teuthology.orchestra.run.vm09.stdout:VERSION="9" 2026-03-10T05:48:50.633 INFO:teuthology.orchestra.run.vm09.stdout:ID="centos" 2026-03-10T05:48:50.633 INFO:teuthology.orchestra.run.vm09.stdout:ID_LIKE="rhel fedora" 2026-03-10T05:48:50.633 INFO:teuthology.orchestra.run.vm09.stdout:VERSION_ID="9" 2026-03-10T05:48:50.633 INFO:teuthology.orchestra.run.vm09.stdout:PLATFORM_ID="platform:el9" 2026-03-10T05:48:50.633 INFO:teuthology.orchestra.run.vm09.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-10T05:48:50.633 INFO:teuthology.orchestra.run.vm09.stdout:ANSI_COLOR="0;31" 2026-03-10T05:48:50.633 INFO:teuthology.orchestra.run.vm09.stdout:LOGO="fedora-logo-icon" 2026-03-10T05:48:50.633 INFO:teuthology.orchestra.run.vm09.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-10T05:48:50.633 INFO:teuthology.orchestra.run.vm09.stdout:HOME_URL="https://centos.org/" 2026-03-10T05:48:50.633 INFO:teuthology.orchestra.run.vm09.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-10T05:48:50.633 INFO:teuthology.orchestra.run.vm09.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-10T05:48:50.633 INFO:teuthology.orchestra.run.vm09.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-10T05:48:50.634 INFO:teuthology.lock.ops:Updating vm09.local on lock server 2026-03-10T05:48:50.638 INFO:teuthology.run_tasks:Running task internal.serialize_remote_roles... 2026-03-10T05:48:50.641 INFO:teuthology.run_tasks:Running task internal.check_conflict... 2026-03-10T05:48:50.642 INFO:teuthology.task.internal:Checking for old test directory... 2026-03-10T05:48:50.642 DEBUG:teuthology.orchestra.run.vm01:> test '!' -e /home/ubuntu/cephtest 2026-03-10T05:48:50.644 DEBUG:teuthology.orchestra.run.vm09:> test '!' -e /home/ubuntu/cephtest 2026-03-10T05:48:50.687 INFO:teuthology.run_tasks:Running task internal.check_ceph_data... 2026-03-10T05:48:50.689 INFO:teuthology.task.internal:Checking for non-empty /var/lib/ceph... 2026-03-10T05:48:50.689 DEBUG:teuthology.orchestra.run.vm01:> test -z $(ls -A /var/lib/ceph) 2026-03-10T05:48:50.697 DEBUG:teuthology.orchestra.run.vm09:> test -z $(ls -A /var/lib/ceph) 2026-03-10T05:48:50.712 INFO:teuthology.orchestra.run.vm01.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-10T05:48:50.743 INFO:teuthology.orchestra.run.vm09.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-10T05:48:50.743 INFO:teuthology.run_tasks:Running task internal.vm_setup... 2026-03-10T05:48:50.753 DEBUG:teuthology.orchestra.run.vm01:> test -e /ceph-qa-ready 2026-03-10T05:48:50.769 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T05:48:50.979 DEBUG:teuthology.orchestra.run.vm09:> test -e /ceph-qa-ready 2026-03-10T05:48:50.996 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T05:48:51.207 INFO:teuthology.run_tasks:Running task internal.base... 2026-03-10T05:48:51.209 INFO:teuthology.task.internal:Creating test directory... 2026-03-10T05:48:51.209 DEBUG:teuthology.orchestra.run.vm01:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-10T05:48:51.211 DEBUG:teuthology.orchestra.run.vm09:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-10T05:48:51.231 INFO:teuthology.run_tasks:Running task internal.archive_upload... 2026-03-10T05:48:51.232 INFO:teuthology.run_tasks:Running task internal.archive... 2026-03-10T05:48:51.234 INFO:teuthology.task.internal:Creating archive directory... 2026-03-10T05:48:51.234 DEBUG:teuthology.orchestra.run.vm01:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-10T05:48:51.269 DEBUG:teuthology.orchestra.run.vm09:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-10T05:48:51.289 INFO:teuthology.run_tasks:Running task internal.coredump... 2026-03-10T05:48:51.290 INFO:teuthology.task.internal:Enabling coredump saving... 2026-03-10T05:48:51.290 DEBUG:teuthology.orchestra.run.vm01:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-10T05:48:51.337 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T05:48:51.337 DEBUG:teuthology.orchestra.run.vm09:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-10T05:48:51.354 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T05:48:51.354 DEBUG:teuthology.orchestra.run.vm01:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-10T05:48:51.379 DEBUG:teuthology.orchestra.run.vm09:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-10T05:48:51.402 INFO:teuthology.orchestra.run.vm01.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T05:48:51.414 INFO:teuthology.orchestra.run.vm01.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T05:48:51.427 INFO:teuthology.orchestra.run.vm09.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T05:48:51.441 INFO:teuthology.orchestra.run.vm09.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T05:48:51.442 INFO:teuthology.run_tasks:Running task internal.sudo... 2026-03-10T05:48:51.445 INFO:teuthology.task.internal:Configuring sudo... 2026-03-10T05:48:51.445 DEBUG:teuthology.orchestra.run.vm01:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-10T05:48:51.458 DEBUG:teuthology.orchestra.run.vm09:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-10T05:48:51.513 INFO:teuthology.run_tasks:Running task internal.syslog... 2026-03-10T05:48:51.515 INFO:teuthology.task.internal.syslog:Starting syslog monitoring... 2026-03-10T05:48:51.516 DEBUG:teuthology.orchestra.run.vm01:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-10T05:48:51.528 DEBUG:teuthology.orchestra.run.vm09:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-10T05:48:51.572 DEBUG:teuthology.orchestra.run.vm01:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T05:48:51.611 DEBUG:teuthology.orchestra.run.vm01:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T05:48:51.667 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T05:48:51.667 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-10T05:48:51.727 DEBUG:teuthology.orchestra.run.vm09:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T05:48:51.754 DEBUG:teuthology.orchestra.run.vm09:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T05:48:51.813 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T05:48:51.813 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-10T05:48:51.871 DEBUG:teuthology.orchestra.run.vm01:> sudo service rsyslog restart 2026-03-10T05:48:51.874 DEBUG:teuthology.orchestra.run.vm09:> sudo service rsyslog restart 2026-03-10T05:48:51.907 INFO:teuthology.orchestra.run.vm01.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-10T05:48:51.939 INFO:teuthology.orchestra.run.vm09.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-10T05:48:52.399 INFO:teuthology.run_tasks:Running task internal.timer... 2026-03-10T05:48:52.401 INFO:teuthology.task.internal:Starting timer... 2026-03-10T05:48:52.401 INFO:teuthology.run_tasks:Running task pcp... 2026-03-10T05:48:52.404 INFO:teuthology.run_tasks:Running task selinux... 2026-03-10T05:48:52.406 DEBUG:teuthology.task:Applying overrides for task selinux: {'allowlist': ['scontext=system_u:system_r:logrotate_t:s0']} 2026-03-10T05:48:52.406 INFO:teuthology.task.selinux:Excluding vm01: VMs are not yet supported 2026-03-10T05:48:52.406 INFO:teuthology.task.selinux:Excluding vm09: VMs are not yet supported 2026-03-10T05:48:52.406 DEBUG:teuthology.task.selinux:Getting current SELinux state 2026-03-10T05:48:52.406 DEBUG:teuthology.task.selinux:Existing SELinux modes: {} 2026-03-10T05:48:52.406 INFO:teuthology.task.selinux:Putting SELinux into permissive mode 2026-03-10T05:48:52.406 INFO:teuthology.run_tasks:Running task ansible.cephlab... 2026-03-10T05:48:52.408 DEBUG:teuthology.task:Applying overrides for task ansible.cephlab: {'branch': 'main', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'timezone': 'UTC'}} 2026-03-10T05:48:52.408 DEBUG:teuthology.repo_utils:Setting repo remote to https://github.com/ceph/ceph-cm-ansible.git 2026-03-10T05:48:52.409 INFO:teuthology.repo_utils:Fetching github.com_ceph_ceph-cm-ansible_main from origin 2026-03-10T05:48:52.968 DEBUG:teuthology.repo_utils:Resetting repo at /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main to origin/main 2026-03-10T05:48:52.974 INFO:teuthology.task.ansible:Playbook: [{'import_playbook': 'ansible_managed.yml'}, {'import_playbook': 'teuthology.yml'}, {'hosts': 'testnodes', 'tasks': [{'set_fact': {'ran_from_cephlab_playbook': True}}]}, {'import_playbook': 'testnodes.yml'}, {'import_playbook': 'container-host.yml'}, {'import_playbook': 'cobbler.yml'}, {'import_playbook': 'paddles.yml'}, {'import_playbook': 'pulpito.yml'}, {'hosts': 'testnodes', 'become': True, 'tasks': [{'name': 'Touch /ceph-qa-ready', 'file': {'path': '/ceph-qa-ready', 'state': 'touch'}, 'when': 'ran_from_cephlab_playbook|bool'}]}] 2026-03-10T05:48:52.974 DEBUG:teuthology.task.ansible:Running ansible-playbook -v --extra-vars '{"ansible_ssh_user": "ubuntu", "timezone": "UTC"}' -i /tmp/teuth_ansible_inventoryqat1e6n7 --limit vm01.local,vm09.local /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main/cephlab.yml --skip-tags nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs 2026-03-10T05:50:31.730 DEBUG:teuthology.task.ansible:Reconnecting to [Remote(name='ubuntu@vm01.local'), Remote(name='ubuntu@vm09.local')] 2026-03-10T05:50:31.730 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm01.local' 2026-03-10T05:50:31.731 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm01.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T05:50:31.799 DEBUG:teuthology.orchestra.run.vm01:> true 2026-03-10T05:50:31.880 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm01.local' 2026-03-10T05:50:31.880 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm09.local' 2026-03-10T05:50:31.880 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm09.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T05:50:31.942 DEBUG:teuthology.orchestra.run.vm09:> true 2026-03-10T05:50:32.022 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm09.local' 2026-03-10T05:50:32.022 INFO:teuthology.run_tasks:Running task clock... 2026-03-10T05:50:32.026 INFO:teuthology.task.clock:Syncing clocks and checking initial clock skew... 2026-03-10T05:50:32.027 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-10T05:50:32.027 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T05:50:32.028 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-10T05:50:32.028 DEBUG:teuthology.orchestra.run.vm09:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T05:50:32.064 INFO:teuthology.orchestra.run.vm01.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-10T05:50:32.086 INFO:teuthology.orchestra.run.vm01.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-10T05:50:32.102 INFO:teuthology.orchestra.run.vm09.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-10T05:50:32.117 INFO:teuthology.orchestra.run.vm01.stderr:sudo: ntpd: command not found 2026-03-10T05:50:32.125 INFO:teuthology.orchestra.run.vm09.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-10T05:50:32.135 INFO:teuthology.orchestra.run.vm01.stdout:506 Cannot talk to daemon 2026-03-10T05:50:32.152 INFO:teuthology.orchestra.run.vm09.stderr:sudo: ntpd: command not found 2026-03-10T05:50:32.153 INFO:teuthology.orchestra.run.vm01.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-10T05:50:32.163 INFO:teuthology.orchestra.run.vm09.stdout:506 Cannot talk to daemon 2026-03-10T05:50:32.166 INFO:teuthology.orchestra.run.vm01.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-10T05:50:32.177 INFO:teuthology.orchestra.run.vm09.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-10T05:50:32.193 INFO:teuthology.orchestra.run.vm09.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-10T05:50:32.218 INFO:teuthology.orchestra.run.vm01.stderr:bash: line 1: ntpq: command not found 2026-03-10T05:50:32.223 INFO:teuthology.orchestra.run.vm01.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-10T05:50:32.223 INFO:teuthology.orchestra.run.vm01.stdout:=============================================================================== 2026-03-10T05:50:32.223 INFO:teuthology.orchestra.run.vm01.stdout:^? mail.trexler.at 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T05:50:32.223 INFO:teuthology.orchestra.run.vm01.stdout:^? a.ntp.madduck.net 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T05:50:32.223 INFO:teuthology.orchestra.run.vm01.stdout:^? static.222.16.42.77.clie> 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T05:50:32.223 INFO:teuthology.orchestra.run.vm01.stdout:^? node-3.infogral.is 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T05:50:32.247 INFO:teuthology.orchestra.run.vm09.stderr:bash: line 1: ntpq: command not found 2026-03-10T05:50:32.251 INFO:teuthology.orchestra.run.vm09.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-10T05:50:32.251 INFO:teuthology.orchestra.run.vm09.stdout:=============================================================================== 2026-03-10T05:50:32.251 INFO:teuthology.orchestra.run.vm09.stdout:^? static.222.16.42.77.clie> 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T05:50:32.251 INFO:teuthology.orchestra.run.vm09.stdout:^? node-3.infogral.is 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T05:50:32.251 INFO:teuthology.orchestra.run.vm09.stdout:^? mail.trexler.at 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T05:50:32.251 INFO:teuthology.orchestra.run.vm09.stdout:^? a.ntp.madduck.net 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T05:50:32.251 INFO:teuthology.run_tasks:Running task pexec... 2026-03-10T05:50:32.255 INFO:teuthology.task.pexec:Executing custom commands... 2026-03-10T05:50:32.255 DEBUG:teuthology.orchestra.run.vm01:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-10T05:50:32.255 DEBUG:teuthology.orchestra.run.vm09:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-10T05:50:32.265 DEBUG:teuthology.task.pexec:ubuntu@vm01.local< sudo dnf remove nvme-cli -y 2026-03-10T05:50:32.266 DEBUG:teuthology.task.pexec:ubuntu@vm01.local< sudo dnf install runc nvmetcli nvme-cli -y 2026-03-10T05:50:32.266 DEBUG:teuthology.task.pexec:ubuntu@vm01.local< sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-10T05:50:32.266 DEBUG:teuthology.task.pexec:ubuntu@vm01.local< sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-10T05:50:32.266 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm01.local 2026-03-10T05:50:32.266 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-10T05:50:32.266 INFO:teuthology.task.pexec:sudo dnf install runc nvmetcli nvme-cli -y 2026-03-10T05:50:32.266 INFO:teuthology.task.pexec:sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-10T05:50:32.266 INFO:teuthology.task.pexec:sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-10T05:50:32.294 DEBUG:teuthology.task.pexec:ubuntu@vm09.local< sudo dnf remove nvme-cli -y 2026-03-10T05:50:32.294 DEBUG:teuthology.task.pexec:ubuntu@vm09.local< sudo dnf install runc nvmetcli nvme-cli -y 2026-03-10T05:50:32.294 DEBUG:teuthology.task.pexec:ubuntu@vm09.local< sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-10T05:50:32.294 DEBUG:teuthology.task.pexec:ubuntu@vm09.local< sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-10T05:50:32.295 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm09.local 2026-03-10T05:50:32.295 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-10T05:50:32.295 INFO:teuthology.task.pexec:sudo dnf install runc nvmetcli nvme-cli -y 2026-03-10T05:50:32.295 INFO:teuthology.task.pexec:sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-10T05:50:32.295 INFO:teuthology.task.pexec:sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-10T05:50:32.495 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: nvme-cli 2026-03-10T05:50:32.495 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-10T05:50:32.498 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:50:32.499 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-10T05:50:32.499 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:50:32.549 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: nvme-cli 2026-03-10T05:50:32.550 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T05:50:32.554 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T05:50:32.554 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T05:50:32.554 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T05:50:32.969 INFO:teuthology.orchestra.run.vm01.stdout:Last metadata expiration check: 0:01:03 ago on Tue 10 Mar 2026 05:49:29 AM UTC. 2026-03-10T05:50:33.039 INFO:teuthology.orchestra.run.vm09.stdout:Last metadata expiration check: 0:01:06 ago on Tue 10 Mar 2026 05:49:27 AM UTC. 2026-03-10T05:50:33.095 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:50:33.095 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:50:33.095 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-10T05:50:33.095 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:50:33.095 INFO:teuthology.orchestra.run.vm01.stdout:Installing: 2026-03-10T05:50:33.095 INFO:teuthology.orchestra.run.vm01.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-10T05:50:33.095 INFO:teuthology.orchestra.run.vm01.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-10T05:50:33.095 INFO:teuthology.orchestra.run.vm01.stdout: runc x86_64 4:1.4.0-2.el9 appstream 4.0 M 2026-03-10T05:50:33.095 INFO:teuthology.orchestra.run.vm01.stdout:Installing dependencies: 2026-03-10T05:50:33.095 INFO:teuthology.orchestra.run.vm01.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-10T05:50:33.095 INFO:teuthology.orchestra.run.vm01.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-10T05:50:33.095 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-10T05:50:33.095 INFO:teuthology.orchestra.run.vm01.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-10T05:50:33.095 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:50:33.095 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-10T05:50:33.095 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:50:33.095 INFO:teuthology.orchestra.run.vm01.stdout:Install 7 Packages 2026-03-10T05:50:33.096 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:50:33.096 INFO:teuthology.orchestra.run.vm01.stdout:Total download size: 6.3 M 2026-03-10T05:50:33.096 INFO:teuthology.orchestra.run.vm01.stdout:Installed size: 24 M 2026-03-10T05:50:33.096 INFO:teuthology.orchestra.run.vm01.stdout:Downloading Packages: 2026-03-10T05:50:33.165 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T05:50:33.166 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T05:50:33.166 INFO:teuthology.orchestra.run.vm09.stdout: Package Arch Version Repository Size 2026-03-10T05:50:33.166 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T05:50:33.166 INFO:teuthology.orchestra.run.vm09.stdout:Installing: 2026-03-10T05:50:33.166 INFO:teuthology.orchestra.run.vm09.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-10T05:50:33.166 INFO:teuthology.orchestra.run.vm09.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-10T05:50:33.166 INFO:teuthology.orchestra.run.vm09.stdout: runc x86_64 4:1.4.0-2.el9 appstream 4.0 M 2026-03-10T05:50:33.166 INFO:teuthology.orchestra.run.vm09.stdout:Installing dependencies: 2026-03-10T05:50:33.166 INFO:teuthology.orchestra.run.vm09.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-10T05:50:33.166 INFO:teuthology.orchestra.run.vm09.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-10T05:50:33.166 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-10T05:50:33.166 INFO:teuthology.orchestra.run.vm09.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-10T05:50:33.166 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T05:50:33.166 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-03-10T05:50:33.166 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T05:50:33.166 INFO:teuthology.orchestra.run.vm09.stdout:Install 7 Packages 2026-03-10T05:50:33.166 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T05:50:33.166 INFO:teuthology.orchestra.run.vm09.stdout:Total download size: 6.3 M 2026-03-10T05:50:33.166 INFO:teuthology.orchestra.run.vm09.stdout:Installed size: 24 M 2026-03-10T05:50:33.166 INFO:teuthology.orchestra.run.vm09.stdout:Downloading Packages: 2026-03-10T05:50:33.520 INFO:teuthology.orchestra.run.vm01.stdout:(1/7): nvmetcli-0.8-3.el9.noarch.rpm 235 kB/s | 44 kB 00:00 2026-03-10T05:50:33.529 INFO:teuthology.orchestra.run.vm01.stdout:(2/7): python3-configshell-1.1.30-1.el9.noarch. 366 kB/s | 72 kB 00:00 2026-03-10T05:50:33.587 INFO:teuthology.orchestra.run.vm01.stdout:(3/7): python3-pyparsing-2.4.7-9.el9.noarch.rpm 2.5 MB/s | 150 kB 00:00 2026-03-10T05:50:33.595 INFO:teuthology.orchestra.run.vm01.stdout:(4/7): python3-kmod-0.9-32.el9.x86_64.rpm 1.1 MB/s | 84 kB 00:00 2026-03-10T05:50:33.640 INFO:teuthology.orchestra.run.vm01.stdout:(5/7): nvme-cli-2.16-1.el9.x86_64.rpm 3.8 MB/s | 1.2 MB 00:00 2026-03-10T05:50:33.681 INFO:teuthology.orchestra.run.vm01.stdout:(6/7): python3-urwid-2.1.2-4.el9.x86_64.rpm 8.8 MB/s | 837 kB 00:00 2026-03-10T05:50:33.854 INFO:teuthology.orchestra.run.vm01.stdout:(7/7): runc-1.4.0-2.el9.x86_64.rpm 15 MB/s | 4.0 MB 00:00 2026-03-10T05:50:33.855 INFO:teuthology.orchestra.run.vm01.stdout:-------------------------------------------------------------------------------- 2026-03-10T05:50:33.855 INFO:teuthology.orchestra.run.vm01.stdout:Total 8.3 MB/s | 6.3 MB 00:00 2026-03-10T05:50:33.927 INFO:teuthology.orchestra.run.vm09.stdout:(1/7): nvmetcli-0.8-3.el9.noarch.rpm 364 kB/s | 44 kB 00:00 2026-03-10T05:50:33.928 INFO:teuthology.orchestra.run.vm09.stdout:(2/7): python3-configshell-1.1.30-1.el9.noarch. 589 kB/s | 72 kB 00:00 2026-03-10T05:50:33.930 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-10T05:50:33.940 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-10T05:50:33.941 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-10T05:50:33.987 INFO:teuthology.orchestra.run.vm09.stdout:(3/7): python3-kmod-0.9-32.el9.x86_64.rpm 1.4 MB/s | 84 kB 00:00 2026-03-10T05:50:33.990 INFO:teuthology.orchestra.run.vm09.stdout:(4/7): python3-pyparsing-2.4.7-9.el9.noarch.rpm 2.4 MB/s | 150 kB 00:00 2026-03-10T05:50:34.015 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-10T05:50:34.015 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-10T05:50:34.112 INFO:teuthology.orchestra.run.vm09.stdout:(5/7): python3-urwid-2.1.2-4.el9.x86_64.rpm 6.5 MB/s | 837 kB 00:00 2026-03-10T05:50:34.220 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-10T05:50:34.232 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/7 2026-03-10T05:50:34.245 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/7 2026-03-10T05:50:34.252 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/7 2026-03-10T05:50:34.259 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/7 2026-03-10T05:50:34.261 INFO:teuthology.orchestra.run.vm01.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/7 2026-03-10T05:50:34.328 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/7 2026-03-10T05:50:34.358 INFO:teuthology.orchestra.run.vm09.stdout:(6/7): nvme-cli-2.16-1.el9.x86_64.rpm 2.1 MB/s | 1.2 MB 00:00 2026-03-10T05:50:34.490 INFO:teuthology.orchestra.run.vm01.stdout: Installing : runc-4:1.4.0-2.el9.x86_64 6/7 2026-03-10T05:50:34.495 INFO:teuthology.orchestra.run.vm01.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 7/7 2026-03-10T05:50:34.914 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 7/7 2026-03-10T05:50:34.914 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-10T05:50:34.914 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:50:35.200 INFO:teuthology.orchestra.run.vm09.stdout:(7/7): runc-1.4.0-2.el9.x86_64.rpm 3.3 MB/s | 4.0 MB 00:01 2026-03-10T05:50:35.200 INFO:teuthology.orchestra.run.vm09.stdout:-------------------------------------------------------------------------------- 2026-03-10T05:50:35.200 INFO:teuthology.orchestra.run.vm09.stdout:Total 3.1 MB/s | 6.3 MB 00:02 2026-03-10T05:50:35.304 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-03-10T05:50:35.312 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-03-10T05:50:35.312 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-03-10T05:50:35.389 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-03-10T05:50:35.389 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-03-10T05:50:35.499 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/7 2026-03-10T05:50:35.500 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/7 2026-03-10T05:50:35.500 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/7 2026-03-10T05:50:35.500 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/7 2026-03-10T05:50:35.500 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/7 2026-03-10T05:50:35.500 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/7 2026-03-10T05:50:35.596 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : runc-4:1.4.0-2.el9.x86_64 7/7 2026-03-10T05:50:35.596 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:50:35.596 INFO:teuthology.orchestra.run.vm01.stdout:Installed: 2026-03-10T05:50:35.596 INFO:teuthology.orchestra.run.vm01.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-10T05:50:35.596 INFO:teuthology.orchestra.run.vm01.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-10T05:50:35.596 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-10T05:50:35.596 INFO:teuthology.orchestra.run.vm01.stdout: runc-4:1.4.0-2.el9.x86_64 2026-03-10T05:50:35.597 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:50:35.597 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:50:35.634 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-03-10T05:50:35.649 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/7 2026-03-10T05:50:35.660 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/7 2026-03-10T05:50:35.671 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/7 2026-03-10T05:50:35.682 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/7 2026-03-10T05:50:35.684 INFO:teuthology.orchestra.run.vm09.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/7 2026-03-10T05:50:35.722 DEBUG:teuthology.parallel:result is None 2026-03-10T05:50:35.749 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/7 2026-03-10T05:50:35.937 INFO:teuthology.orchestra.run.vm09.stdout: Installing : runc-4:1.4.0-2.el9.x86_64 6/7 2026-03-10T05:50:35.950 INFO:teuthology.orchestra.run.vm09.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 7/7 2026-03-10T05:50:36.396 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 7/7 2026-03-10T05:50:36.397 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-10T05:50:36.397 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T05:50:37.015 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/7 2026-03-10T05:50:37.015 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/7 2026-03-10T05:50:37.015 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/7 2026-03-10T05:50:37.015 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/7 2026-03-10T05:50:37.015 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/7 2026-03-10T05:50:37.015 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/7 2026-03-10T05:50:37.122 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : runc-4:1.4.0-2.el9.x86_64 7/7 2026-03-10T05:50:37.122 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T05:50:37.122 INFO:teuthology.orchestra.run.vm09.stdout:Installed: 2026-03-10T05:50:37.122 INFO:teuthology.orchestra.run.vm09.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-10T05:50:37.122 INFO:teuthology.orchestra.run.vm09.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-10T05:50:37.122 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-10T05:50:37.122 INFO:teuthology.orchestra.run.vm09.stdout: runc-4:1.4.0-2.el9.x86_64 2026-03-10T05:50:37.122 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T05:50:37.122 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T05:50:37.237 DEBUG:teuthology.parallel:result is None 2026-03-10T05:50:37.237 INFO:teuthology.run_tasks:Running task cephadm... 2026-03-10T05:50:37.285 INFO:tasks.cephadm:Config: {'roleless': True, 'conf': {'mgr': {'debug mgr': 20, 'debug ms': 1}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000, 'osd shutdown pgref assert': True}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', 'CEPHADM_DAEMON_PLACE_FAIL', 'CEPHADM_FAILED_DAEMON'], 'log-only-match': ['CEPHADM_'], 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'} 2026-03-10T05:50:37.285 INFO:tasks.cephadm:Cluster image is quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T05:50:37.285 INFO:tasks.cephadm:Cluster fsid is 100dc3f2-1c45-11f1-8010-b96ca19c2df4 2026-03-10T05:50:37.285 INFO:tasks.cephadm:Choosing monitor IPs and ports... 2026-03-10T05:50:37.285 INFO:tasks.cephadm:No mon roles; fabricating mons 2026-03-10T05:50:37.285 INFO:tasks.cephadm:Monitor IPs: {'mon.vm01': '192.168.123.101', 'mon.vm09': '192.168.123.109'} 2026-03-10T05:50:37.285 INFO:tasks.cephadm:Normalizing hostnames... 2026-03-10T05:50:37.285 DEBUG:teuthology.orchestra.run.vm01:> sudo hostname $(hostname -s) 2026-03-10T05:50:37.327 DEBUG:teuthology.orchestra.run.vm09:> sudo hostname $(hostname -s) 2026-03-10T05:50:37.369 INFO:tasks.cephadm:Downloading "compiled" cephadm from cachra 2026-03-10T05:50:37.369 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T05:50:38.043 INFO:tasks.cephadm:builder_project result: [{'url': 'https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/', 'chacra_url': 'https://3.chacra.ceph.com/repos/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/', 'ref': 'squid', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'distro': 'centos', 'distro_version': '9', 'distro_codename': None, 'modified': '2026-02-25 18:55:15.146628', 'status': 'ready', 'flavor': 'default', 'project': 'ceph', 'archs': ['source', 'x86_64'], 'extra': {'version': '19.2.3-678-ge911bdeb', 'package_manager_version': '19.2.3-678.ge911bdeb', 'build_url': 'https://jenkins.ceph.com/job/ceph-dev-pipeline/3275/', 'root_build_cause': '', 'node_name': '10.20.192.26+soko16', 'job_name': 'ceph-dev-pipeline'}}] 2026-03-10T05:50:38.639 INFO:tasks.util.chacra:got chacra host 3.chacra.ceph.com, ref squid, sha1 e911bdebe5c8faa3800735d1568fcdca65db60df from https://shaman.ceph.com/api/search/?project=ceph&distros=centos%2F9%2Fx86_64&flavor=default&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T05:50:38.640 INFO:tasks.cephadm:Discovered cachra url: https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm 2026-03-10T05:50:38.640 INFO:tasks.cephadm:Downloading cephadm from url: https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm 2026-03-10T05:50:38.640 DEBUG:teuthology.orchestra.run.vm01:> curl --silent -L https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-10T05:50:40.142 INFO:teuthology.orchestra.run.vm01.stdout:-rw-r--r--. 1 ubuntu ubuntu 788355 Mar 10 05:50 /home/ubuntu/cephtest/cephadm 2026-03-10T05:50:40.142 DEBUG:teuthology.orchestra.run.vm09:> curl --silent -L https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-10T05:50:41.659 INFO:teuthology.orchestra.run.vm09.stdout:-rw-r--r--. 1 ubuntu ubuntu 788355 Mar 10 05:50 /home/ubuntu/cephtest/cephadm 2026-03-10T05:50:41.659 DEBUG:teuthology.orchestra.run.vm01:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-10T05:50:41.680 DEBUG:teuthology.orchestra.run.vm09:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-10T05:50:41.702 INFO:tasks.cephadm:Pulling image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df on all hosts... 2026-03-10T05:50:41.703 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-10T05:50:41.723 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-10T05:50:41.913 INFO:teuthology.orchestra.run.vm01.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T05:50:41.972 INFO:teuthology.orchestra.run.vm09.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T05:51:50.402 INFO:teuthology.orchestra.run.vm09.stdout:{ 2026-03-10T05:51:50.403 INFO:teuthology.orchestra.run.vm09.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-10T05:51:50.403 INFO:teuthology.orchestra.run.vm09.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-10T05:51:50.403 INFO:teuthology.orchestra.run.vm09.stdout: "repo_digests": [ 2026-03-10T05:51:50.403 INFO:teuthology.orchestra.run.vm09.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-10T05:51:50.403 INFO:teuthology.orchestra.run.vm09.stdout: ] 2026-03-10T05:51:50.403 INFO:teuthology.orchestra.run.vm09.stdout:} 2026-03-10T05:52:36.342 INFO:teuthology.orchestra.run.vm01.stdout:{ 2026-03-10T05:52:36.343 INFO:teuthology.orchestra.run.vm01.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-10T05:52:36.343 INFO:teuthology.orchestra.run.vm01.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-10T05:52:36.343 INFO:teuthology.orchestra.run.vm01.stdout: "repo_digests": [ 2026-03-10T05:52:36.343 INFO:teuthology.orchestra.run.vm01.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-10T05:52:36.343 INFO:teuthology.orchestra.run.vm01.stdout: ] 2026-03-10T05:52:36.343 INFO:teuthology.orchestra.run.vm01.stdout:} 2026-03-10T05:52:36.360 DEBUG:teuthology.orchestra.run.vm01:> sudo mkdir -p /etc/ceph 2026-03-10T05:52:36.385 DEBUG:teuthology.orchestra.run.vm09:> sudo mkdir -p /etc/ceph 2026-03-10T05:52:36.411 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod 777 /etc/ceph 2026-03-10T05:52:36.449 DEBUG:teuthology.orchestra.run.vm09:> sudo chmod 777 /etc/ceph 2026-03-10T05:52:36.476 INFO:tasks.cephadm:Writing seed config... 2026-03-10T05:52:36.477 INFO:tasks.cephadm: override: [mgr] debug mgr = 20 2026-03-10T05:52:36.477 INFO:tasks.cephadm: override: [mgr] debug ms = 1 2026-03-10T05:52:36.477 INFO:tasks.cephadm: override: [mon] debug mon = 20 2026-03-10T05:52:36.477 INFO:tasks.cephadm: override: [mon] debug ms = 1 2026-03-10T05:52:36.477 INFO:tasks.cephadm: override: [mon] debug paxos = 20 2026-03-10T05:52:36.477 INFO:tasks.cephadm: override: [osd] debug ms = 1 2026-03-10T05:52:36.477 INFO:tasks.cephadm: override: [osd] debug osd = 20 2026-03-10T05:52:36.477 INFO:tasks.cephadm: override: [osd] osd mclock iops capacity threshold hdd = 49000 2026-03-10T05:52:36.477 INFO:tasks.cephadm: override: [osd] osd shutdown pgref assert = True 2026-03-10T05:52:36.477 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T05:52:36.477 DEBUG:teuthology.orchestra.run.vm01:> dd of=/home/ubuntu/cephtest/seed.ceph.conf 2026-03-10T05:52:36.504 DEBUG:tasks.cephadm:Final config: [global] # make logging friendly to teuthology log_to_file = true log_to_stderr = false log to journald = false mon cluster log to file = true mon cluster log file level = debug mon clock drift allowed = 1.000 # replicate across OSDs, not hosts osd crush chooseleaf type = 0 #osd pool default size = 2 osd pool default erasure code profile = plugin=jerasure technique=reed_sol_van k=2 m=1 crush-failure-domain=osd # enable some debugging auth debug = true ms die on old message = true ms die on bug = true debug asserts on shutdown = true # adjust warnings mon max pg per osd = 10000# >= luminous mon pg warn max object skew = 0 mon osd allow primary affinity = true mon osd allow pg remap = true mon warn on legacy crush tunables = false mon warn on crush straw calc version zero = false mon warn on no sortbitwise = false mon warn on osd down out interval zero = false mon warn on too few osds = false mon_warn_on_pool_pg_num_not_power_of_two = false # disable pg_autoscaler by default for new pools osd_pool_default_pg_autoscale_mode = off # tests delete pools mon allow pool delete = true fsid = 100dc3f2-1c45-11f1-8010-b96ca19c2df4 [osd] osd scrub load threshold = 5.0 osd scrub max interval = 600 osd mclock profile = high_recovery_ops osd recover clone overlap = true osd recovery max chunk = 1048576 osd deep scrub update digest min age = 30 osd map max advance = 10 osd memory target autotune = true # debugging osd debug shutdown = true osd debug op order = true osd debug verify stray on activate = true osd debug pg log writeout = true osd debug verify cached snaps = true osd debug verify missing on start = true osd debug misdirected ops = true osd op queue = debug_random osd op queue cut off = debug_random osd shutdown pgref assert = True bdev debug aio = true osd sloppy crc = true debug ms = 1 debug osd = 20 osd mclock iops capacity threshold hdd = 49000 [mgr] mon reweight min pgs per osd = 4 mon reweight min bytes per osd = 10 mgr/telemetry/nag = false debug mgr = 20 debug ms = 1 [mon] mon data avail warn = 5 mon mgr mkfs grace = 240 mon reweight min pgs per osd = 4 mon osd reporter subtree level = osd mon osd prime pg temp = true mon reweight min bytes per osd = 10 # rotate auth tickets quickly to exercise renewal paths auth mon ticket ttl = 660# 11m auth service ticket ttl = 240# 4m # don't complain about global id reclaim mon_warn_on_insecure_global_id_reclaim = false mon_warn_on_insecure_global_id_reclaim_allowed = false debug mon = 20 debug ms = 1 debug paxos = 20 [client.rgw] rgw cache enabled = true rgw enable ops log = true rgw enable usage log = true 2026-03-10T05:52:36.504 DEBUG:teuthology.orchestra.run.vm01:mon.vm01> sudo journalctl -f -n 0 -u ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4@mon.vm01.service 2026-03-10T05:52:36.546 INFO:tasks.cephadm:Bootstrapping... 2026-03-10T05:52:36.546 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df -v bootstrap --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 --config /home/ubuntu/cephtest/seed.ceph.conf --output-config /etc/ceph/ceph.conf --output-keyring /etc/ceph/ceph.client.admin.keyring --output-pub-ssh-key /home/ubuntu/cephtest/ceph.pub --mon-ip 192.168.123.101 --skip-admin-label && sudo chmod +r /etc/ceph/ceph.client.admin.keyring 2026-03-10T05:52:36.685 INFO:teuthology.orchestra.run.vm01.stdout:-------------------------------------------------------------------------------- 2026-03-10T05:52:36.685 INFO:teuthology.orchestra.run.vm01.stdout:cephadm ['--image', 'quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df', '-v', 'bootstrap', '--fsid', '100dc3f2-1c45-11f1-8010-b96ca19c2df4', '--config', '/home/ubuntu/cephtest/seed.ceph.conf', '--output-config', '/etc/ceph/ceph.conf', '--output-keyring', '/etc/ceph/ceph.client.admin.keyring', '--output-pub-ssh-key', '/home/ubuntu/cephtest/ceph.pub', '--mon-ip', '192.168.123.101', '--skip-admin-label'] 2026-03-10T05:52:36.685 INFO:teuthology.orchestra.run.vm01.stderr:Specifying an fsid for your cluster offers no advantages and may increase the likelihood of fsid conflicts. 2026-03-10T05:52:36.685 INFO:teuthology.orchestra.run.vm01.stdout:Verifying podman|docker is present... 2026-03-10T05:52:36.704 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stdout 5.8.0 2026-03-10T05:52:36.704 INFO:teuthology.orchestra.run.vm01.stdout:Verifying lvm2 is present... 2026-03-10T05:52:36.704 INFO:teuthology.orchestra.run.vm01.stdout:Verifying time synchronization is in place... 2026-03-10T05:52:36.710 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-10T05:52:36.710 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-10T05:52:36.715 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-10T05:52:36.715 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout inactive 2026-03-10T05:52:36.721 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout enabled 2026-03-10T05:52:36.726 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout active 2026-03-10T05:52:36.727 INFO:teuthology.orchestra.run.vm01.stdout:Unit chronyd.service is enabled and running 2026-03-10T05:52:36.727 INFO:teuthology.orchestra.run.vm01.stdout:Repeating the final host check... 2026-03-10T05:52:36.745 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stdout 5.8.0 2026-03-10T05:52:36.745 INFO:teuthology.orchestra.run.vm01.stdout:podman (/bin/podman) version 5.8.0 is present 2026-03-10T05:52:36.745 INFO:teuthology.orchestra.run.vm01.stdout:systemctl is present 2026-03-10T05:52:36.745 INFO:teuthology.orchestra.run.vm01.stdout:lvcreate is present 2026-03-10T05:52:36.751 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-10T05:52:36.751 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-10T05:52:36.757 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-10T05:52:36.757 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout inactive 2026-03-10T05:52:36.764 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout enabled 2026-03-10T05:52:36.770 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout active 2026-03-10T05:52:36.770 INFO:teuthology.orchestra.run.vm01.stdout:Unit chronyd.service is enabled and running 2026-03-10T05:52:36.770 INFO:teuthology.orchestra.run.vm01.stdout:Host looks OK 2026-03-10T05:52:36.770 INFO:teuthology.orchestra.run.vm01.stdout:Cluster fsid: 100dc3f2-1c45-11f1-8010-b96ca19c2df4 2026-03-10T05:52:36.770 INFO:teuthology.orchestra.run.vm01.stdout:Acquiring lock 139874019582304 on /run/cephadm/100dc3f2-1c45-11f1-8010-b96ca19c2df4.lock 2026-03-10T05:52:36.770 INFO:teuthology.orchestra.run.vm01.stdout:Lock 139874019582304 acquired on /run/cephadm/100dc3f2-1c45-11f1-8010-b96ca19c2df4.lock 2026-03-10T05:52:36.771 INFO:teuthology.orchestra.run.vm01.stdout:Verifying IP 192.168.123.101 port 3300 ... 2026-03-10T05:52:36.771 INFO:teuthology.orchestra.run.vm01.stdout:Verifying IP 192.168.123.101 port 6789 ... 2026-03-10T05:52:36.771 INFO:teuthology.orchestra.run.vm01.stdout:Base mon IP(s) is [192.168.123.101:3300, 192.168.123.101:6789], mon addrv is [v2:192.168.123.101:3300,v1:192.168.123.101:6789] 2026-03-10T05:52:36.774 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout default via 192.168.123.1 dev eth0 proto dhcp src 192.168.123.101 metric 100 2026-03-10T05:52:36.774 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout 192.168.123.0/24 dev eth0 proto kernel scope link src 192.168.123.101 metric 100 2026-03-10T05:52:36.777 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout ::1 dev lo proto kernel metric 256 pref medium 2026-03-10T05:52:36.777 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout fe80::/64 dev eth0 proto kernel metric 1024 pref medium 2026-03-10T05:52:36.780 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout 1: lo: mtu 65536 state UNKNOWN qlen 1000 2026-03-10T05:52:36.780 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout inet6 ::1/128 scope host 2026-03-10T05:52:36.780 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-10T05:52:36.780 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout 2: eth0: mtu 1500 state UP qlen 1000 2026-03-10T05:52:36.780 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout inet6 fe80::5055:ff:fe00:1/64 scope link noprefixroute 2026-03-10T05:52:36.780 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-10T05:52:36.780 INFO:teuthology.orchestra.run.vm01.stdout:Mon IP `192.168.123.101` is in CIDR network `192.168.123.0/24` 2026-03-10T05:52:36.781 INFO:teuthology.orchestra.run.vm01.stdout:Mon IP `192.168.123.101` is in CIDR network `192.168.123.0/24` 2026-03-10T05:52:36.781 INFO:teuthology.orchestra.run.vm01.stdout:Inferred mon public CIDR from local network configuration ['192.168.123.0/24', '192.168.123.0/24'] 2026-03-10T05:52:36.781 INFO:teuthology.orchestra.run.vm01.stdout:Internal network (--cluster-network) has not been provided, OSD replication will default to the public_network 2026-03-10T05:52:36.781 INFO:teuthology.orchestra.run.vm01.stdout:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T05:52:38.110 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stdout 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c 2026-03-10T05:52:38.110 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Trying to pull quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T05:52:38.110 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Getting image source signatures 2026-03-10T05:52:38.110 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Copying blob sha256:1752b8d01aa0dd33bbe0ab24e8316174c94fbdcd5d26252e2680bba0624747a7 2026-03-10T05:52:38.110 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Copying blob sha256:8e380faede39ebd4286247457b408d979ab568aafd8389c42ec304b8cfba4e92 2026-03-10T05:52:38.110 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Copying config sha256:654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c 2026-03-10T05:52:38.110 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Writing manifest to image destination 2026-03-10T05:52:38.374 INFO:teuthology.orchestra.run.vm01.stdout:ceph: stdout ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable) 2026-03-10T05:52:38.374 INFO:teuthology.orchestra.run.vm01.stdout:Ceph version: ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable) 2026-03-10T05:52:38.374 INFO:teuthology.orchestra.run.vm01.stdout:Extracting ceph user uid/gid from container image... 2026-03-10T05:52:38.574 INFO:teuthology.orchestra.run.vm01.stdout:stat: stdout 167 167 2026-03-10T05:52:38.575 INFO:teuthology.orchestra.run.vm01.stdout:Creating initial keys... 2026-03-10T05:52:38.910 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-authtool: stdout AQCmsa9ptQ6hKBAAsZhksNnwChvO+45wULYucQ== 2026-03-10T05:52:39.227 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-authtool: stdout AQCnsa9pR1d4ARAAn9rZRBc1nb0BUwSG64o1rQ== 2026-03-10T05:52:39.517 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-authtool: stdout AQCnsa9pXvYWFBAAmzEF8/pBORlrXH0DgF19kQ== 2026-03-10T05:52:39.517 INFO:teuthology.orchestra.run.vm01.stdout:Creating initial monmap... 2026-03-10T05:52:39.965 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-10T05:52:39.965 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: stdout setting min_mon_release = quincy 2026-03-10T05:52:39.965 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: set fsid to 100dc3f2-1c45-11f1-8010-b96ca19c2df4 2026-03-10T05:52:39.965 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-10T05:52:39.965 INFO:teuthology.orchestra.run.vm01.stdout:monmaptool for vm01 [v2:192.168.123.101:3300,v1:192.168.123.101:6789] on /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-10T05:52:39.966 INFO:teuthology.orchestra.run.vm01.stdout:setting min_mon_release = quincy 2026-03-10T05:52:39.966 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: set fsid to 100dc3f2-1c45-11f1-8010-b96ca19c2df4 2026-03-10T05:52:39.966 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-10T05:52:39.966 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:52:39.966 INFO:teuthology.orchestra.run.vm01.stdout:Creating mon... 2026-03-10T05:52:40.359 INFO:teuthology.orchestra.run.vm01.stdout:create mon.vm01 on 2026-03-10T05:52:40.603 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /etc/systemd/system/ceph.target. 2026-03-10T05:52:40.714 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4.target → /etc/systemd/system/ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4.target. 2026-03-10T05:52:40.714 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph.target.wants/ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4.target → /etc/systemd/system/ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4.target. 2026-03-10T05:52:40.858 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4@mon.vm01 2026-03-10T05:52:40.858 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Failed to reset failed state of unit ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4@mon.vm01.service: Unit ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4@mon.vm01.service not loaded. 2026-03-10T05:52:40.988 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4.target.wants/ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4@mon.vm01.service → /etc/systemd/system/ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4@.service. 2026-03-10T05:52:41.131 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:41 vm01 podman[46872]: 2026-03-10 05:52:41.100203833 +0000 UTC m=+0.016156879 container create dcb28a6cf37a5e9ae76e09f7ffa7a17e4bc1b0b33e0d208b72c0c09f7de6c4aa (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4-mon-vm01, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223, org.label-schema.license=GPLv2, org.opencontainers.image.authors=Ceph Release Team , ceph=True, org.label-schema.schema-version=1.0, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.name=CentOS Stream 9 Base Image, FROM_IMAGE=quay.io/centos/centos:stream9, io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.vendor=CentOS, CEPH_REF=squid, OSD_FLAVOR=default, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git) 2026-03-10T05:52:41.146 INFO:teuthology.orchestra.run.vm01.stdout:firewalld does not appear to be present 2026-03-10T05:52:41.147 INFO:teuthology.orchestra.run.vm01.stdout:Not possible to enable service . firewalld.service is not available 2026-03-10T05:52:41.147 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mon to start... 2026-03-10T05:52:41.147 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mon... 2026-03-10T05:52:41.457 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:41 vm01 podman[46872]: 2026-03-10 05:52:41.134233281 +0000 UTC m=+0.050186327 container init dcb28a6cf37a5e9ae76e09f7ffa7a17e4bc1b0b33e0d208b72c0c09f7de6c4aa (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4-mon-vm01, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, OSD_FLAVOR=default, io.buildah.version=1.41.3, CEPH_REF=squid, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.build-date=20260223, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.authors=Ceph Release Team , ceph=True, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS) 2026-03-10T05:52:41.457 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:41 vm01 podman[46872]: 2026-03-10 05:52:41.137971397 +0000 UTC m=+0.053924443 container start dcb28a6cf37a5e9ae76e09f7ffa7a17e4bc1b0b33e0d208b72c0c09f7de6c4aa (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4-mon-vm01, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.build-date=20260223, CEPH_REF=squid, OSD_FLAVOR=default, FROM_IMAGE=quay.io/centos/centos:stream9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True) 2026-03-10T05:52:41.457 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:41 vm01 bash[46872]: dcb28a6cf37a5e9ae76e09f7ffa7a17e4bc1b0b33e0d208b72c0c09f7de6c4aa 2026-03-10T05:52:41.457 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:41 vm01 podman[46872]: 2026-03-10 05:52:41.093585587 +0000 UTC m=+0.009538642 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T05:52:41.457 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:41 vm01 systemd[1]: Started Ceph mon.vm01 for 100dc3f2-1c45-11f1-8010-b96ca19c2df4. 2026-03-10T05:52:41.457 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:41 vm01 ceph-mon[46908]: mkfs 100dc3f2-1c45-11f1-8010-b96ca19c2df4 2026-03-10T05:52:41.457 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:41 vm01 ceph-mon[46908]: mon.vm01 is new leader, mons vm01 in quorum (ranks 0) 2026-03-10T05:52:41.465 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout cluster: 2026-03-10T05:52:41.465 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout id: 100dc3f2-1c45-11f1-8010-b96ca19c2df4 2026-03-10T05:52:41.465 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout health: HEALTH_OK 2026-03-10T05:52:41.465 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T05:52:41.465 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout services: 2026-03-10T05:52:41.465 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon: 1 daemons, quorum vm01 (age 0.141882s) 2026-03-10T05:52:41.465 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mgr: no daemons active 2026-03-10T05:52:41.465 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd: 0 osds: 0 up, 0 in 2026-03-10T05:52:41.465 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T05:52:41.465 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout data: 2026-03-10T05:52:41.465 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout pools: 0 pools, 0 pgs 2026-03-10T05:52:41.465 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout objects: 0 objects, 0 B 2026-03-10T05:52:41.465 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout usage: 0 B used, 0 B / 0 B avail 2026-03-10T05:52:41.465 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout pgs: 2026-03-10T05:52:41.465 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T05:52:41.465 INFO:teuthology.orchestra.run.vm01.stdout:mon is available 2026-03-10T05:52:41.465 INFO:teuthology.orchestra.run.vm01.stdout:Assimilating anything we can from ceph.conf... 2026-03-10T05:52:41.772 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T05:52:41.772 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [global] 2026-03-10T05:52:41.772 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout fsid = 100dc3f2-1c45-11f1-8010-b96ca19c2df4 2026-03-10T05:52:41.772 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-10T05:52:41.772 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.101:3300,v1:192.168.123.101:6789] 2026-03-10T05:52:41.772 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-10T05:52:41.772 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-10T05:52:41.772 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-10T05:52:41.772 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-10T05:52:41.772 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T05:52:41.772 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-10T05:52:41.772 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-10T05:52:41.772 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T05:52:41.772 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [osd] 2026-03-10T05:52:41.772 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-10T05:52:41.772 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-10T05:52:41.772 INFO:teuthology.orchestra.run.vm01.stdout:Generating new minimal ceph.conf... 2026-03-10T05:52:42.068 INFO:teuthology.orchestra.run.vm01.stdout:Restarting the monitor... 2026-03-10T05:52:42.325 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 systemd[1]: Stopping Ceph mon.vm01 for 100dc3f2-1c45-11f1-8010-b96ca19c2df4... 2026-03-10T05:52:42.325 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4-mon-vm01[46884]: 2026-03-10T05:52:42.140+0000 7f2b7131f640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.vm01 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T05:52:42.325 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4-mon-vm01[46884]: 2026-03-10T05:52:42.140+0000 7f2b7131f640 -1 mon.vm01@0(leader) e1 *** Got Signal Terminated *** 2026-03-10T05:52:42.585 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 podman[47184]: 2026-03-10 05:52:42.325408989 +0000 UTC m=+0.198426424 container died dcb28a6cf37a5e9ae76e09f7ffa7a17e4bc1b0b33e0d208b72c0c09f7de6c4aa (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4-mon-vm01, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, org.label-schema.build-date=20260223, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, OSD_FLAVOR=default, CEPH_REF=squid, io.buildah.version=1.41.3, org.opencontainers.image.authors=Ceph Release Team ) 2026-03-10T05:52:42.585 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 podman[47184]: 2026-03-10 05:52:42.443920238 +0000 UTC m=+0.316937673 container remove dcb28a6cf37a5e9ae76e09f7ffa7a17e4bc1b0b33e0d208b72c0c09f7de6c4aa (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4-mon-vm01, OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_REF=squid, org.label-schema.build-date=20260223, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-10T05:52:42.585 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 bash[47184]: ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4-mon-vm01 2026-03-10T05:52:42.585 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 systemd[1]: ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4@mon.vm01.service: Deactivated successfully. 2026-03-10T05:52:42.585 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 systemd[1]: Stopped Ceph mon.vm01 for 100dc3f2-1c45-11f1-8010-b96ca19c2df4. 2026-03-10T05:52:42.585 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 systemd[1]: Starting Ceph mon.vm01 for 100dc3f2-1c45-11f1-8010-b96ca19c2df4... 2026-03-10T05:52:42.633 INFO:teuthology.orchestra.run.vm01.stdout:Setting public_network to 192.168.123.0/24 in mon config section 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 podman[47263]: 2026-03-10 05:52:42.585682833 +0000 UTC m=+0.016091918 container create d0fd780adbeca7225f8c248f65412d9a3e15a7cd74b9bcd9cf0de6f323e725f7 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4-mon-vm01, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, OSD_FLAVOR=default, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20260223, org.opencontainers.image.documentation=https://docs.ceph.com/, org.opencontainers.image.authors=Ceph Release Team , ceph=True, org.label-schema.vendor=CentOS) 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 podman[47263]: 2026-03-10 05:52:42.619678237 +0000 UTC m=+0.050087332 container init d0fd780adbeca7225f8c248f65412d9a3e15a7cd74b9bcd9cf0de6f323e725f7 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4-mon-vm01, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True, org.opencontainers.image.documentation=https://docs.ceph.com/, io.buildah.version=1.41.3, CEPH_REF=squid, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS, org.opencontainers.image.authors=Ceph Release Team , CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 podman[47263]: 2026-03-10 05:52:42.623603914 +0000 UTC m=+0.054012998 container start d0fd780adbeca7225f8c248f65412d9a3e15a7cd74b9bcd9cf0de6f323e725f7 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4-mon-vm01, OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.vendor=CentOS, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.authors=Ceph Release Team , io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, CEPH_REF=squid, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, ceph=True, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 bash[47263]: d0fd780adbeca7225f8c248f65412d9a3e15a7cd74b9bcd9cf0de6f323e725f7 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 podman[47263]: 2026-03-10 05:52:42.579756282 +0000 UTC m=+0.010165366 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 systemd[1]: Started Ceph mon.vm01 for 100dc3f2-1c45-11f1-8010-b96ca19c2df4. 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: set uid:gid to 167:167 (ceph:ceph) 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable), process ceph-mon, pid 7 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: pidfile_write: ignore empty --pid-file 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: load: jerasure load: lrc 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: RocksDB version: 7.9.2 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Git sha 0 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Compile date 2026-02-25 18:11:04 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: DB SUMMARY 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: DB Session ID: 1CZUSM8C16YOOTLN986X 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: CURRENT file: CURRENT 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: IDENTITY file: IDENTITY 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: MANIFEST file: MANIFEST-000010 size: 179 Bytes 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: SST files in /var/lib/ceph/mon/ceph-vm01/store.db dir, Total Num: 1, files: 000008.sst 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-vm01/store.db: 000009.log size: 75099 ; 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.error_if_exists: 0 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.create_if_missing: 0 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.paranoid_checks: 1 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.flush_verify_memtable_count: 1 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.env: 0x55ef32070dc0 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.fs: PosixFileSystem 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.info_log: 0x55ef33ec4b20 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.max_file_opening_threads: 16 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.statistics: (nil) 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.use_fsync: 0 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.max_log_file_size: 0 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.log_file_time_to_roll: 0 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.keep_log_file_num: 1000 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.recycle_log_file_num: 0 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.allow_fallocate: 1 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.allow_mmap_reads: 0 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.allow_mmap_writes: 0 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.use_direct_reads: 0 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.create_missing_column_families: 0 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.db_log_dir: 2026-03-10T05:52:42.945 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.wal_dir: 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.table_cache_numshardbits: 6 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.WAL_ttl_seconds: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.WAL_size_limit_MB: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.is_fd_close_on_exec: 1 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.advise_random_on_open: 1 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.db_write_buffer_size: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.write_buffer_manager: 0x55ef33ec9900 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.use_adaptive_mutex: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.rate_limiter: (nil) 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.wal_recovery_mode: 2 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.enable_thread_tracking: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.enable_pipelined_write: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.unordered_write: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.row_cache: None 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.wal_filter: None 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.allow_ingest_behind: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.two_write_queues: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.manual_wal_flush: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.wal_compression: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.atomic_flush: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.persist_stats_to_disk: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.write_dbid_to_manifest: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.log_readahead_size: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.best_efforts_recovery: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.allow_data_in_errors: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.db_host_id: __hostname__ 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.enforce_single_del_contracts: true 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.max_background_jobs: 2 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.max_background_compactions: -1 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.max_subcompactions: 1 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.delayed_write_rate : 16777216 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.max_total_wal_size: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.stats_dump_period_sec: 600 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.stats_persist_period_sec: 600 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.max_open_files: -1 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.bytes_per_sync: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.wal_bytes_per_sync: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.strict_bytes_per_sync: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.compaction_readahead_size: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.max_background_flushes: -1 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Compression algorithms supported: 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: kZSTD supported: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: kXpressCompression supported: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: kBZip2Compression supported: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-10T05:52:42.946 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: kLZ4Compression supported: 1 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: kZlibCompression supported: 1 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: kLZ4HCCompression supported: 1 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: kSnappyCompression supported: 1 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Fast CRC32 supported: Supported on x86 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: DMutex implementation: pthread_mutex_t 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-vm01/store.db/MANIFEST-000010 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.merge_operator: 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.compaction_filter: None 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.compaction_filter_factory: None 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.sst_partitioner_factory: None 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.memtable_factory: SkipListFactory 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.table_factory: BlockBasedTable 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55ef33ec46e0) 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: cache_index_and_filter_blocks: 1 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: pin_top_level_index_and_filter: 1 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: index_type: 0 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: data_block_index_type: 0 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: index_shortening: 1 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: data_block_hash_table_util_ratio: 0.750000 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: checksum: 4 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: no_block_cache: 0 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: block_cache: 0x55ef33ee9350 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: block_cache_name: BinnedLRUCache 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: block_cache_options: 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: capacity : 536870912 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: num_shard_bits : 4 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: strict_capacity_limit : 0 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: high_pri_pool_ratio: 0.000 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: block_cache_compressed: (nil) 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: persistent_cache: (nil) 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: block_size: 4096 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: block_size_deviation: 10 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: block_restart_interval: 16 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: index_block_restart_interval: 1 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: metadata_block_size: 4096 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: partition_filters: 0 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: use_delta_encoding: 1 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: filter_policy: bloomfilter 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: whole_key_filtering: 1 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: verify_compression: 0 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: read_amp_bytes_per_bit: 0 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: format_version: 5 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: enable_index_compression: 1 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: block_align: 0 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: max_auto_readahead_size: 262144 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: prepopulate_block_cache: 0 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: initial_auto_readahead_size: 8192 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout: num_file_reads_for_auto_readahead: 2 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.write_buffer_size: 33554432 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.max_write_buffer_number: 2 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.compression: NoCompression 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.bottommost_compression: Disabled 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.prefix_extractor: nullptr 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.num_levels: 7 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-10T05:52:42.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.compression_opts.window_bits: -14 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.compression_opts.level: 32767 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.compression_opts.strategy: 0 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.compression_opts.enabled: false 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.target_file_size_base: 67108864 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.target_file_size_multiplier: 1 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.arena_block_size: 1048576 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.disable_auto_compactions: 0 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.inplace_update_support: 0 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.inplace_update_num_locks: 10000 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.memtable_huge_page_size: 0 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.bloom_locality: 0 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.max_successive_merges: 0 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.optimize_filters_for_hits: 0 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.paranoid_file_checks: 0 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.force_consistency_checks: 1 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.report_bg_io_stats: 0 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.ttl: 2592000 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.periodic_compaction_seconds: 0 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.enable_blob_files: false 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.min_blob_size: 0 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.blob_file_size: 268435456 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.blob_compression_type: NoCompression 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.enable_blob_garbage_collection: false 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-10T05:52:42.948 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.blob_file_starting_level: 0 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-vm01/store.db/MANIFEST-000010 succeeded,manifest_file_number is 10, next_file_number is 12, last_sequence is 5, log_number is 5,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 5 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 5 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: a4b7daca-6edd-4abc-a84f-7f7138392016 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773121962648865, "job": 1, "event": "recovery_started", "wal_files": [9]} 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #9 mode 2 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773121962650325, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 13, "file_size": 72167, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 8, "largest_seqno": 223, "table_properties": {"data_size": 70446, "index_size": 174, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 517, "raw_key_size": 9562, "raw_average_key_size": 49, "raw_value_size": 65071, "raw_average_value_size": 335, "num_data_blocks": 8, "num_entries": 194, "num_filter_entries": 194, "num_deletions": 3, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1773121962, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "a4b7daca-6edd-4abc-a84f-7f7138392016", "db_session_id": "1CZUSM8C16YOOTLN986X", "orig_file_number": 13, "seqno_to_time_mapping": "N/A"}} 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773121962650388, "job": 1, "event": "recovery_finished"} 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: [db/version_set.cc:5047] Creating manifest 15 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-vm01/store.db/000009.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55ef33eeae00 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: DB pointer 0x55ef34004000 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: ** DB Stats ** 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: ** Compaction Stats [default] ** 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: L0 2/0 72.35 KB 0.5 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 54.2 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: Sum 2/0 72.35 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 54.2 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 54.2 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: ** Compaction Stats [default] ** 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 54.2 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: Flush(GB): cumulative 0.000, interval 0.000 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: AddFile(Total Files): cumulative 0, interval 0 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: AddFile(L0 Files): cumulative 0, interval 0 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: AddFile(Keys): cumulative 0, interval 0 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: Cumulative compaction: 0.00 GB write, 12.16 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: Interval compaction: 0.00 GB write, 12.16 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: Block cache BinnedLRUCache@0x55ef33ee9350#7 capacity: 512.00 MB usage: 2.02 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 2.8e-05 secs_since: 0 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: Block cache entry stats(count,size,portion): FilterBlock(2,0.70 KB,0.00013411%) IndexBlock(2,0.36 KB,6.85453e-05%) Misc(2,0.95 KB,0.000181794%) 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: ** File Read Latency Histogram By Level [default] ** 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: starting mon.vm01 rank 0 at public addrs [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] at bind addrs [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon_data /var/lib/ceph/mon/ceph-vm01 fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: mon.vm01@-1(???) e1 preinit fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: mon.vm01@-1(???).mds e1 new map 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: mon.vm01@-1(???).mds e1 print_map 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: e1 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: btime 2026-03-10T05:52:41:179707+0000 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: enable_multiple, ever_enabled_multiple: 1,1 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: legacy client fscid: -1 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout: No filesystems configured 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: mon.vm01@-1(???).osd e1 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: mon.vm01@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T05:52:42.949 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: mon.vm01@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T05:52:42.950 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: mon.vm01@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T05:52:42.950 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: mon.vm01@-1(???).paxosservice(auth 1..2) refresh upgraded, format 0 -> 3 2026-03-10T05:52:42.950 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: mon.vm01 is new leader, mons vm01 in quorum (ranks 0) 2026-03-10T05:52:42.950 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: monmap epoch 1 2026-03-10T05:52:42.950 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 2026-03-10T05:52:42.950 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: last_changed 2026-03-10T05:52:39.621470+0000 2026-03-10T05:52:42.950 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: created 2026-03-10T05:52:39.621470+0000 2026-03-10T05:52:42.950 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: min_mon_release 19 (squid) 2026-03-10T05:52:42.950 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: election_strategy: 1 2026-03-10T05:52:42.950 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.vm01 2026-03-10T05:52:42.950 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: fsmap 2026-03-10T05:52:42.950 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: osdmap e1: 0 total, 0 up, 0 in 2026-03-10T05:52:42.950 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:42 vm01 ceph-mon[47299]: mgrmap e1: no daemons active 2026-03-10T05:52:42.950 INFO:teuthology.orchestra.run.vm01.stdout:Wrote config to /etc/ceph/ceph.conf 2026-03-10T05:52:42.950 INFO:teuthology.orchestra.run.vm01.stdout:Wrote keyring to /etc/ceph/ceph.client.admin.keyring 2026-03-10T05:52:42.950 INFO:teuthology.orchestra.run.vm01.stdout:Creating mgr... 2026-03-10T05:52:42.950 INFO:teuthology.orchestra.run.vm01.stdout:Verifying port 0.0.0.0:9283 ... 2026-03-10T05:52:42.951 INFO:teuthology.orchestra.run.vm01.stdout:Verifying port 0.0.0.0:8765 ... 2026-03-10T05:52:42.951 INFO:teuthology.orchestra.run.vm01.stdout:Verifying port 0.0.0.0:8443 ... 2026-03-10T05:52:43.097 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4@mgr.vm01.wforik 2026-03-10T05:52:43.097 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Failed to reset failed state of unit ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4@mgr.vm01.wforik.service: Unit ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4@mgr.vm01.wforik.service not loaded. 2026-03-10T05:52:43.223 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4.target.wants/ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4@mgr.vm01.wforik.service → /etc/systemd/system/ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4@.service. 2026-03-10T05:52:43.389 INFO:teuthology.orchestra.run.vm01.stdout:firewalld does not appear to be present 2026-03-10T05:52:43.389 INFO:teuthology.orchestra.run.vm01.stdout:Not possible to enable service . firewalld.service is not available 2026-03-10T05:52:43.389 INFO:teuthology.orchestra.run.vm01.stdout:firewalld does not appear to be present 2026-03-10T05:52:43.389 INFO:teuthology.orchestra.run.vm01.stdout:Not possible to open ports <[9283, 8765, 8443]>. firewalld.service is not available 2026-03-10T05:52:43.389 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mgr to start... 2026-03-10T05:52:43.389 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mgr... 2026-03-10T05:52:43.755 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T05:52:43.755 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-10T05:52:43.755 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsid": "100dc3f2-1c45-11f1-8010-b96ca19c2df4", 2026-03-10T05:52:43.755 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "health": { 2026-03-10T05:52:43.755 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-10T05:52:43.755 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-10T05:52:43.755 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-10T05:52:43.755 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:52:43.755 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-10T05:52:43.755 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-10T05:52:43.755 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 0 2026-03-10T05:52:43.755 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-10T05:52:43.755 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-10T05:52:43.755 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "vm01" 2026-03-10T05:52:43.755 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-10T05:52:43.755 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_age": 0, 2026-03-10T05:52:43.755 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-10T05:52:43.755 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T05:52:43.755 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-10T05:52:43.755 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-10T05:52:43.755 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:52:43.755 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-10T05:52:43.755 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T05:52:43.755 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-10T05:52:43.755 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-10T05:52:43.755 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-10T05:52:43.755 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-10T05:52:43.755 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "btime": "2026-03-10T05:52:41:179707+0000", 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "restful" 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modified": "2026-03-10T05:52:41.180311+0000", 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-10T05:52:43.756 INFO:teuthology.orchestra.run.vm01.stdout:mgr not available, waiting (1/15)... 2026-03-10T05:52:44.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:43 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/3186782452' entity='client.admin' 2026-03-10T05:52:44.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:43 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/2171342023' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T05:52:46.112 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:46 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/344241115' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T05:52:46.126 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T05:52:46.126 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-10T05:52:46.126 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsid": "100dc3f2-1c45-11f1-8010-b96ca19c2df4", 2026-03-10T05:52:46.126 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "health": { 2026-03-10T05:52:46.126 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-10T05:52:46.126 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-10T05:52:46.126 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-10T05:52:46.126 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:52:46.126 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-10T05:52:46.126 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-10T05:52:46.126 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 0 2026-03-10T05:52:46.126 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-10T05:52:46.126 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-10T05:52:46.126 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "vm01" 2026-03-10T05:52:46.126 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-10T05:52:46.126 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_age": 3, 2026-03-10T05:52:46.126 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "btime": "2026-03-10T05:52:41:179707+0000", 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "restful" 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modified": "2026-03-10T05:52:41.180311+0000", 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-10T05:52:46.127 INFO:teuthology.orchestra.run.vm01.stdout:mgr not available, waiting (2/15)... 2026-03-10T05:52:47.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:47 vm01 ceph-mon[47299]: Activating manager daemon vm01.wforik 2026-03-10T05:52:47.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:47 vm01 ceph-mon[47299]: mgrmap e2: vm01.wforik(active, starting, since 0.0038552s) 2026-03-10T05:52:47.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:47 vm01 ceph-mon[47299]: from='mgr.14100 192.168.123.101:0/1126556124' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-10T05:52:47.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:47 vm01 ceph-mon[47299]: from='mgr.14100 192.168.123.101:0/1126556124' entity='mgr.vm01.wforik' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T05:52:47.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:47 vm01 ceph-mon[47299]: from='mgr.14100 192.168.123.101:0/1126556124' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T05:52:47.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:47 vm01 ceph-mon[47299]: from='mgr.14100 192.168.123.101:0/1126556124' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T05:52:47.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:47 vm01 ceph-mon[47299]: from='mgr.14100 192.168.123.101:0/1126556124' entity='mgr.vm01.wforik' cmd=[{"prefix": "mgr metadata", "who": "vm01.wforik", "id": "vm01.wforik"}]: dispatch 2026-03-10T05:52:47.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:47 vm01 ceph-mon[47299]: Manager daemon vm01.wforik is now available 2026-03-10T05:52:47.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:47 vm01 ceph-mon[47299]: from='mgr.14100 192.168.123.101:0/1126556124' entity='mgr.vm01.wforik' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.wforik/mirror_snapshot_schedule"}]: dispatch 2026-03-10T05:52:47.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:47 vm01 ceph-mon[47299]: from='mgr.14100 192.168.123.101:0/1126556124' entity='mgr.vm01.wforik' 2026-03-10T05:52:47.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:47 vm01 ceph-mon[47299]: from='mgr.14100 192.168.123.101:0/1126556124' entity='mgr.vm01.wforik' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.wforik/trash_purge_schedule"}]: dispatch 2026-03-10T05:52:47.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:47 vm01 ceph-mon[47299]: from='mgr.14100 192.168.123.101:0/1126556124' entity='mgr.vm01.wforik' 2026-03-10T05:52:47.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:47 vm01 ceph-mon[47299]: from='mgr.14100 192.168.123.101:0/1126556124' entity='mgr.vm01.wforik' 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsid": "100dc3f2-1c45-11f1-8010-b96ca19c2df4", 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "health": { 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 0 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "vm01" 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_age": 5, 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-10T05:52:48.781 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "btime": "2026-03-10T05:52:41:179707+0000", 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "restful" 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modified": "2026-03-10T05:52:41.180311+0000", 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-10T05:52:48.782 INFO:teuthology.orchestra.run.vm01.stdout:mgr is available 2026-03-10T05:52:49.177 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T05:52:49.177 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [global] 2026-03-10T05:52:49.177 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout fsid = 100dc3f2-1c45-11f1-8010-b96ca19c2df4 2026-03-10T05:52:49.177 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-10T05:52:49.177 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.101:3300,v1:192.168.123.101:6789] 2026-03-10T05:52:49.177 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-10T05:52:49.177 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-10T05:52:49.177 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-10T05:52:49.178 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-10T05:52:49.178 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T05:52:49.178 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-10T05:52:49.178 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-10T05:52:49.178 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T05:52:49.178 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [osd] 2026-03-10T05:52:49.178 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-10T05:52:49.178 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-10T05:52:49.178 INFO:teuthology.orchestra.run.vm01.stdout:Enabling cephadm module... 2026-03-10T05:52:49.292 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:49 vm01 ceph-mon[47299]: mgrmap e3: vm01.wforik(active, since 1.83786s) 2026-03-10T05:52:49.292 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:49 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/256650801' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T05:52:49.292 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:49 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/2438874528' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-10T05:52:50.696 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:50 vm01 ceph-mon[47299]: mgrmap e4: vm01.wforik(active, since 2s) 2026-03-10T05:52:50.696 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:50 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/1968105725' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-10T05:52:50.842 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-10T05:52:50.842 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 5, 2026-03-10T05:52:50.842 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-10T05:52:50.843 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "active_name": "vm01.wforik", 2026-03-10T05:52:50.843 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-10T05:52:50.843 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-10T05:52:50.843 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for the mgr to restart... 2026-03-10T05:52:50.843 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mgr epoch 5... 2026-03-10T05:52:52.014 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:51 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/1968105725' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-10T05:52:52.014 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:51 vm01 ceph-mon[47299]: mgrmap e5: vm01.wforik(active, since 3s) 2026-03-10T05:52:52.014 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:51 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/3286629993' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T05:52:53.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:53 vm01 ceph-mon[47299]: Active manager daemon vm01.wforik restarted 2026-03-10T05:52:53.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:53 vm01 ceph-mon[47299]: Activating manager daemon vm01.wforik 2026-03-10T05:52:53.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:53 vm01 ceph-mon[47299]: osdmap e2: 0 total, 0 up, 0 in 2026-03-10T05:52:53.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:53 vm01 ceph-mon[47299]: mgrmap e6: vm01.wforik(active, starting, since 0.00395s) 2026-03-10T05:52:53.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:53 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-10T05:52:53.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:53 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' cmd=[{"prefix": "mgr metadata", "who": "vm01.wforik", "id": "vm01.wforik"}]: dispatch 2026-03-10T05:52:53.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:53 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T05:52:53.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:53 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T05:52:53.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:53 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T05:52:53.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:53 vm01 ceph-mon[47299]: Manager daemon vm01.wforik is now available 2026-03-10T05:52:53.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:53 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:52:53.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:53 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:52:53.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:53 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:52:53.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:53 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.wforik/mirror_snapshot_schedule"}]: dispatch 2026-03-10T05:52:53.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:53 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:52:53.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:53 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.wforik/trash_purge_schedule"}]: dispatch 2026-03-10T05:52:54.568 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-10T05:52:54.568 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 7, 2026-03-10T05:52:54.568 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-10T05:52:54.568 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-10T05:52:54.568 INFO:teuthology.orchestra.run.vm01.stdout:mgr epoch 5 is available 2026-03-10T05:52:54.568 INFO:teuthology.orchestra.run.vm01.stdout:Setting orchestrator backend to cephadm... 2026-03-10T05:52:54.935 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:54 vm01 ceph-mon[47299]: Found migration_current of "None". Setting to last migration. 2026-03-10T05:52:54.936 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:54 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:52:54.936 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:54 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:52:54.936 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:54 vm01 ceph-mon[47299]: mgrmap e7: vm01.wforik(active, since 1.00718s) 2026-03-10T05:52:54.936 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:54 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:52:54.936 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:54 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:52:54.936 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:54 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:52:55.317 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout value unchanged 2026-03-10T05:52:55.317 INFO:teuthology.orchestra.run.vm01.stdout:Generating ssh key... 2026-03-10T05:52:56.081 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:55 vm01 ceph-mon[47299]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-10T05:52:56.081 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:55 vm01 ceph-mon[47299]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-10T05:52:56.081 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:55 vm01 ceph-mon[47299]: [10/Mar/2026:05:52:54] ENGINE Bus STARTING 2026-03-10T05:52:56.081 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:55 vm01 ceph-mon[47299]: [10/Mar/2026:05:52:54] ENGINE Serving on http://192.168.123.101:8765 2026-03-10T05:52:56.081 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:55 vm01 ceph-mon[47299]: [10/Mar/2026:05:52:54] ENGINE Serving on https://192.168.123.101:7150 2026-03-10T05:52:56.081 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:55 vm01 ceph-mon[47299]: [10/Mar/2026:05:52:54] ENGINE Bus STARTED 2026-03-10T05:52:56.081 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:55 vm01 ceph-mon[47299]: [10/Mar/2026:05:52:54] ENGINE Client ('192.168.123.101', 60110) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T05:52:56.081 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:55 vm01 ceph-mon[47299]: from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:52:56.081 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:55 vm01 ceph-mon[47299]: from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:52:56.081 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:55 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:52:56.081 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:55 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:52:56.087 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIDk7QCnyAaqUPpAAIg2+GF3erdEkj09JCRnlCamWsBco ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4 2026-03-10T05:52:56.087 INFO:teuthology.orchestra.run.vm01.stdout:Wrote public SSH key to /home/ubuntu/cephtest/ceph.pub 2026-03-10T05:52:56.087 INFO:teuthology.orchestra.run.vm01.stdout:Adding key to root@localhost authorized_keys... 2026-03-10T05:52:56.087 INFO:teuthology.orchestra.run.vm01.stdout:Adding host vm01... 2026-03-10T05:52:57.530 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:57 vm01 ceph-mon[47299]: from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:52:57.531 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:57 vm01 ceph-mon[47299]: Generating ssh key... 2026-03-10T05:52:57.531 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:57 vm01 ceph-mon[47299]: mgrmap e8: vm01.wforik(active, since 2s) 2026-03-10T05:52:57.531 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:57 vm01 ceph-mon[47299]: from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:52:57.531 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:57 vm01 ceph-mon[47299]: from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm01", "addr": "192.168.123.101", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:52:57.993 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Added host 'vm01' with addr '192.168.123.101' 2026-03-10T05:52:57.994 INFO:teuthology.orchestra.run.vm01.stdout:Deploying mon service with default placement... 2026-03-10T05:52:58.396 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:58 vm01 ceph-mon[47299]: Deploying cephadm binary to vm01 2026-03-10T05:52:58.396 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:58 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:52:58.396 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:58 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:52:58.396 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:58 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:52:58.404 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Scheduled mon update... 2026-03-10T05:52:58.404 INFO:teuthology.orchestra.run.vm01.stdout:Deploying mgr service with default placement... 2026-03-10T05:52:58.769 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Scheduled mgr update... 2026-03-10T05:52:58.769 INFO:teuthology.orchestra.run.vm01.stdout:Deploying crash service with default placement... 2026-03-10T05:52:59.145 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Scheduled crash update... 2026-03-10T05:52:59.145 INFO:teuthology.orchestra.run.vm01.stdout:Deploying ceph-exporter service with default placement... 2026-03-10T05:52:59.511 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:59 vm01 ceph-mon[47299]: Added host vm01 2026-03-10T05:52:59.511 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:59 vm01 ceph-mon[47299]: from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:52:59.511 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:59 vm01 ceph-mon[47299]: Saving service mon spec with placement count:5 2026-03-10T05:52:59.511 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:59 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:52:59.512 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:52:59 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:52:59.519 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Scheduled ceph-exporter update... 2026-03-10T05:52:59.519 INFO:teuthology.orchestra.run.vm01.stdout:Deploying prometheus service with default placement... 2026-03-10T05:52:59.913 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Scheduled prometheus update... 2026-03-10T05:52:59.913 INFO:teuthology.orchestra.run.vm01.stdout:Deploying grafana service with default placement... 2026-03-10T05:53:00.304 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:00 vm01 ceph-mon[47299]: from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:00.304 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:00 vm01 ceph-mon[47299]: Saving service mgr spec with placement count:2 2026-03-10T05:53:00.304 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:00 vm01 ceph-mon[47299]: from='client.14144 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "crash", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:00.304 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:00 vm01 ceph-mon[47299]: Saving service crash spec with placement * 2026-03-10T05:53:00.304 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:00 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:00.304 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:00 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:00.304 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:00 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:00.304 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:00 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:00.304 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:00 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:00.304 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:00 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:00.314 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Scheduled grafana update... 2026-03-10T05:53:00.314 INFO:teuthology.orchestra.run.vm01.stdout:Deploying node-exporter service with default placement... 2026-03-10T05:53:00.723 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Scheduled node-exporter update... 2026-03-10T05:53:00.724 INFO:teuthology.orchestra.run.vm01.stdout:Deploying alertmanager service with default placement... 2026-03-10T05:53:01.131 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Scheduled alertmanager update... 2026-03-10T05:53:01.380 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:01 vm01 ceph-mon[47299]: from='client.14146 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "ceph-exporter", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:01.380 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:01 vm01 ceph-mon[47299]: Saving service ceph-exporter spec with placement * 2026-03-10T05:53:01.380 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:01 vm01 ceph-mon[47299]: from='client.14148 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "prometheus", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:01.380 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:01 vm01 ceph-mon[47299]: Saving service prometheus spec with placement count:1 2026-03-10T05:53:01.380 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:01 vm01 ceph-mon[47299]: from='client.14150 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "grafana", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:01.380 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:01 vm01 ceph-mon[47299]: Saving service grafana spec with placement count:1 2026-03-10T05:53:01.380 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:01 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:01.380 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:01 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:01.380 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:01 vm01 ceph-mon[47299]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:01.867 INFO:teuthology.orchestra.run.vm01.stdout:Enabling the dashboard module... 2026-03-10T05:53:02.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:02 vm01 ceph-mon[47299]: from='client.14152 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "node-exporter", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:02.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:02 vm01 ceph-mon[47299]: Saving service node-exporter spec with placement * 2026-03-10T05:53:02.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:02 vm01 ceph-mon[47299]: from='client.14154 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "alertmanager", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:02.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:02 vm01 ceph-mon[47299]: Saving service alertmanager spec with placement count:1 2026-03-10T05:53:02.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:02 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/2380699250' entity='client.admin' 2026-03-10T05:53:02.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:02 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/3486127958' entity='client.admin' 2026-03-10T05:53:02.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:02 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/1832610603' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-10T05:53:03.286 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-10T05:53:03.286 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 9, 2026-03-10T05:53:03.286 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-10T05:53:03.286 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "active_name": "vm01.wforik", 2026-03-10T05:53:03.286 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-10T05:53:03.286 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-10T05:53:03.286 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for the mgr to restart... 2026-03-10T05:53:03.286 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mgr epoch 9... 2026-03-10T05:53:03.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:03 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/1832610603' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-10T05:53:03.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:03 vm01 ceph-mon[47299]: mgrmap e9: vm01.wforik(active, since 9s) 2026-03-10T05:53:03.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:03 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/4273479382' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T05:53:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:05 vm01 ceph-mon[47299]: Active manager daemon vm01.wforik restarted 2026-03-10T05:53:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:05 vm01 ceph-mon[47299]: Activating manager daemon vm01.wforik 2026-03-10T05:53:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:05 vm01 ceph-mon[47299]: osdmap e3: 0 total, 0 up, 0 in 2026-03-10T05:53:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:05 vm01 ceph-mon[47299]: mgrmap e10: vm01.wforik(active, starting, since 0.00526092s) 2026-03-10T05:53:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:05 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-10T05:53:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:05 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix": "mgr metadata", "who": "vm01.wforik", "id": "vm01.wforik"}]: dispatch 2026-03-10T05:53:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:05 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T05:53:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:05 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T05:53:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:05 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T05:53:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:05 vm01 ceph-mon[47299]: Manager daemon vm01.wforik is now available 2026-03-10T05:53:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:05 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.wforik/mirror_snapshot_schedule"}]: dispatch 2026-03-10T05:53:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:05 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:53:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:05 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.wforik/trash_purge_schedule"}]: dispatch 2026-03-10T05:53:06.992 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-10T05:53:06.993 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 11, 2026-03-10T05:53:06.993 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-10T05:53:06.993 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-10T05:53:06.993 INFO:teuthology.orchestra.run.vm01.stdout:mgr epoch 9 is available 2026-03-10T05:53:06.993 INFO:teuthology.orchestra.run.vm01.stdout:Generating a dashboard self-signed certificate... 2026-03-10T05:53:07.419 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Self-signed certificate created 2026-03-10T05:53:07.420 INFO:teuthology.orchestra.run.vm01.stdout:Creating initial admin user... 2026-03-10T05:53:07.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:07 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:07.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:07 vm01 ceph-mon[47299]: [10/Mar/2026:05:53:06] ENGINE Bus STARTING 2026-03-10T05:53:07.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:07 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:07.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:07 vm01 ceph-mon[47299]: mgrmap e11: vm01.wforik(active, since 1.00872s) 2026-03-10T05:53:07.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:07 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:07.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:07 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:07.953 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout {"username": "admin", "password": "$2b$12$aIf/PqxcROorXxSdUxoH4O.BXG///WFcnTunZKKvSX64X0oiAsg7i", "roles": ["administrator"], "name": null, "email": null, "lastUpdate": 1773121987, "enabled": true, "pwdExpirationDate": null, "pwdUpdateRequired": true} 2026-03-10T05:53:07.953 INFO:teuthology.orchestra.run.vm01.stdout:Fetching dashboard port number... 2026-03-10T05:53:08.341 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 8443 2026-03-10T05:53:08.341 INFO:teuthology.orchestra.run.vm01.stdout:firewalld does not appear to be present 2026-03-10T05:53:08.341 INFO:teuthology.orchestra.run.vm01.stdout:Not possible to open ports <[8443]>. firewalld.service is not available 2026-03-10T05:53:08.341 INFO:teuthology.orchestra.run.vm01.stdout:Ceph Dashboard is now available at: 2026-03-10T05:53:08.341 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:53:08.341 INFO:teuthology.orchestra.run.vm01.stdout: URL: https://vm01.local:8443/ 2026-03-10T05:53:08.341 INFO:teuthology.orchestra.run.vm01.stdout: User: admin 2026-03-10T05:53:08.341 INFO:teuthology.orchestra.run.vm01.stdout: Password: xckdieehyg 2026-03-10T05:53:08.341 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:53:08.341 INFO:teuthology.orchestra.run.vm01.stdout:Saving cluster configuration to /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/config directory 2026-03-10T05:53:08.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:08 vm01 ceph-mon[47299]: from='client.14166 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-10T05:53:08.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:08 vm01 ceph-mon[47299]: from='client.14166 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-10T05:53:08.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:08 vm01 ceph-mon[47299]: [10/Mar/2026:05:53:06] ENGINE Serving on https://192.168.123.101:7150 2026-03-10T05:53:08.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:08 vm01 ceph-mon[47299]: [10/Mar/2026:05:53:06] ENGINE Client ('192.168.123.101', 54832) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T05:53:08.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:08 vm01 ceph-mon[47299]: [10/Mar/2026:05:53:06] ENGINE Serving on http://192.168.123.101:8765 2026-03-10T05:53:08.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:08 vm01 ceph-mon[47299]: [10/Mar/2026:05:53:06] ENGINE Bus STARTED 2026-03-10T05:53:08.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:08 vm01 ceph-mon[47299]: from='client.14174 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:08.619 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:08 vm01 ceph-mon[47299]: from='client.14176 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:08.620 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:08 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:08.620 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:08 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/1548582451' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-10T05:53:08.751 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stderr set mgr/dashboard/cluster/status 2026-03-10T05:53:08.751 INFO:teuthology.orchestra.run.vm01.stdout:You can access the Ceph CLI as following in case of multi-cluster or non-default config: 2026-03-10T05:53:08.751 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:53:08.751 INFO:teuthology.orchestra.run.vm01.stdout: sudo /home/ubuntu/cephtest/cephadm shell --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring 2026-03-10T05:53:08.751 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:53:08.751 INFO:teuthology.orchestra.run.vm01.stdout:Or, if you are only running a single cluster on this host: 2026-03-10T05:53:08.751 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:53:08.751 INFO:teuthology.orchestra.run.vm01.stdout: sudo /home/ubuntu/cephtest/cephadm shell 2026-03-10T05:53:08.751 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:53:08.751 INFO:teuthology.orchestra.run.vm01.stdout:Please consider enabling telemetry to help improve Ceph: 2026-03-10T05:53:08.751 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:53:08.751 INFO:teuthology.orchestra.run.vm01.stdout: ceph telemetry on 2026-03-10T05:53:08.751 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:53:08.751 INFO:teuthology.orchestra.run.vm01.stdout:For more information see: 2026-03-10T05:53:08.751 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:53:08.751 INFO:teuthology.orchestra.run.vm01.stdout: https://docs.ceph.com/en/latest/mgr/telemetry/ 2026-03-10T05:53:08.751 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:53:08.751 INFO:teuthology.orchestra.run.vm01.stdout:Bootstrap complete. 2026-03-10T05:53:08.779 INFO:tasks.cephadm:Fetching config... 2026-03-10T05:53:08.779 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T05:53:08.779 DEBUG:teuthology.orchestra.run.vm01:> dd if=/etc/ceph/ceph.conf of=/dev/stdout 2026-03-10T05:53:08.796 INFO:tasks.cephadm:Fetching client.admin keyring... 2026-03-10T05:53:08.796 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T05:53:08.796 DEBUG:teuthology.orchestra.run.vm01:> dd if=/etc/ceph/ceph.client.admin.keyring of=/dev/stdout 2026-03-10T05:53:08.872 INFO:tasks.cephadm:Fetching mon keyring... 2026-03-10T05:53:08.872 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T05:53:08.872 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/keyring of=/dev/stdout 2026-03-10T05:53:08.946 INFO:tasks.cephadm:Fetching pub ssh key... 2026-03-10T05:53:08.946 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T05:53:08.946 DEBUG:teuthology.orchestra.run.vm01:> dd if=/home/ubuntu/cephtest/ceph.pub of=/dev/stdout 2026-03-10T05:53:09.006 INFO:tasks.cephadm:Installing pub ssh key for root users... 2026-03-10T05:53:09.006 DEBUG:teuthology.orchestra.run.vm01:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIDk7QCnyAaqUPpAAIg2+GF3erdEkj09JCRnlCamWsBco ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-10T05:53:09.086 INFO:teuthology.orchestra.run.vm01.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIDk7QCnyAaqUPpAAIg2+GF3erdEkj09JCRnlCamWsBco ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4 2026-03-10T05:53:09.099 DEBUG:teuthology.orchestra.run.vm09:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIDk7QCnyAaqUPpAAIg2+GF3erdEkj09JCRnlCamWsBco ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-10T05:53:09.134 INFO:teuthology.orchestra.run.vm09.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIDk7QCnyAaqUPpAAIg2+GF3erdEkj09JCRnlCamWsBco ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4 2026-03-10T05:53:09.147 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph config set mgr mgr/cephadm/allow_ptrace true 2026-03-10T05:53:09.344 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:53:09.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:09 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/3276924622' entity='client.admin' 2026-03-10T05:53:09.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:09 vm01 ceph-mon[47299]: mgrmap e12: vm01.wforik(active, since 2s) 2026-03-10T05:53:09.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:09 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/278616664' entity='client.admin' 2026-03-10T05:53:09.780 INFO:tasks.cephadm:Distributing conf and client.admin keyring to all hosts + 0755 2026-03-10T05:53:09.780 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph orch client-keyring set client.admin '*' --mode 0755 2026-03-10T05:53:10.050 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:53:10.541 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm09 2026-03-10T05:53:10.541 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T05:53:10.541 DEBUG:teuthology.orchestra.run.vm09:> dd of=/etc/ceph/ceph.conf 2026-03-10T05:53:10.560 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T05:53:10.560 DEBUG:teuthology.orchestra.run.vm09:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T05:53:10.616 INFO:tasks.cephadm:Adding host vm09 to orchestrator... 2026-03-10T05:53:10.616 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph orch host add vm09 2026-03-10T05:53:10.799 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:53:11.179 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:10 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:11.179 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:10 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:11.179 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:10 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:53:11.179 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:10 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:11.179 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:10 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm01", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-10T05:53:11.179 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:10 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd='[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm01", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]': finished 2026-03-10T05:53:11.179 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:10 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:11.179 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:10 vm01 ceph-mon[47299]: Deploying daemon ceph-exporter.vm01 on vm01 2026-03-10T05:53:11.179 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:10 vm01 ceph-mon[47299]: from='client.14184 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:11.179 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:10 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:12.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:11 vm01 ceph-mon[47299]: from='client.14186 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm09", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:12.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:11 vm01 ceph-mon[47299]: Deploying cephadm binary to vm09 2026-03-10T05:53:12.627 INFO:teuthology.orchestra.run.vm01.stdout:Added host 'vm09' with addr '192.168.123.109' 2026-03-10T05:53:12.862 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph orch host ls --format=json 2026-03-10T05:53:13.099 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:12 vm01 ceph-mon[47299]: mgrmap e13: vm01.wforik(active, since 6s) 2026-03-10T05:53:13.099 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:12 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:13.099 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:12 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:13.099 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:12 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:13.099 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:12 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:13.099 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:12 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm01", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-10T05:53:13.099 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:12 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd='[{"prefix": "auth get-or-create", "entity": "client.crash.vm01", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]': finished 2026-03-10T05:53:13.099 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:12 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:13.099 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:12 vm01 ceph-mon[47299]: Deploying daemon crash.vm01 on vm01 2026-03-10T05:53:13.099 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:12 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:13.099 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:12 vm01 ceph-mon[47299]: Added host vm09 2026-03-10T05:53:13.204 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:53:13.465 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:53:13.466 INFO:teuthology.orchestra.run.vm01.stdout:[{"addr": "192.168.123.101", "hostname": "vm01", "labels": [], "status": ""}, {"addr": "192.168.123.109", "hostname": "vm09", "labels": [], "status": ""}] 2026-03-10T05:53:13.619 INFO:tasks.cephadm:Setting crush tunables to default 2026-03-10T05:53:13.619 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd crush tunables default 2026-03-10T05:53:13.896 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:53:14.213 INFO:teuthology.orchestra.run.vm01.stderr:adjusted tunables profile to default 2026-03-10T05:53:14.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:14 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:14.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:14 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:14.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:14 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:14.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:14 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:14.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:14 vm01 ceph-mon[47299]: Deploying daemon node-exporter.vm01 on vm01 2026-03-10T05:53:14.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:14 vm01 ceph-mon[47299]: from='client.14189 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T05:53:14.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:14 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/3876165708' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-10T05:53:14.380 INFO:tasks.cephadm:Adding mon.vm01 on vm01 2026-03-10T05:53:14.380 INFO:tasks.cephadm:Adding mon.vm09 on vm09 2026-03-10T05:53:14.380 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph orch apply mon '2;vm01:192.168.123.101=vm01;vm09:192.168.123.109=vm09' 2026-03-10T05:53:14.535 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:14.567 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:14.812 INFO:teuthology.orchestra.run.vm09.stdout:Scheduled mon update... 2026-03-10T05:53:14.972 DEBUG:teuthology.orchestra.run.vm09:mon.vm09> sudo journalctl -f -n 0 -u ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4@mon.vm09.service 2026-03-10T05:53:14.974 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T05:53:14.974 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph mon dump -f json 2026-03-10T05:53:15.157 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:15.189 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:15.443 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T05:53:15.443 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"100dc3f2-1c45-11f1-8010-b96ca19c2df4","modified":"2026-03-10T05:52:39.621470Z","created":"2026-03-10T05:52:39.621470Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T05:53:15.444 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T05:53:15.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:15 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/3876165708' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-10T05:53:15.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:15 vm01 ceph-mon[47299]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T05:53:15.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:15 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:16.616 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T05:53:16.616 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph mon dump -f json 2026-03-10T05:53:16.711 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:16 vm01 ceph-mon[47299]: from='client.14193 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "2;vm01:192.168.123.101=vm01;vm09:192.168.123.109=vm09", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:16.711 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:16 vm01 ceph-mon[47299]: Saving service mon spec with placement vm01:192.168.123.101=vm01;vm09:192.168.123.109=vm09;count:2 2026-03-10T05:53:16.711 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:16 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/285557039' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:16.711 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:16 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:16.711 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:16 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:16.711 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:16 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:16.711 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:16 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:16.711 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:16 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:16.782 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:16.811 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:17.085 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T05:53:17.085 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"100dc3f2-1c45-11f1-8010-b96ca19c2df4","modified":"2026-03-10T05:52:39.621470Z","created":"2026-03-10T05:52:39.621470Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T05:53:17.085 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T05:53:17.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:17 vm01 ceph-mon[47299]: Deploying daemon alertmanager.vm01 on vm01 2026-03-10T05:53:17.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:17 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/951122430' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:18.250 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T05:53:18.250 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph mon dump -f json 2026-03-10T05:53:18.410 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:18.442 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:18.799 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T05:53:18.799 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"100dc3f2-1c45-11f1-8010-b96ca19c2df4","modified":"2026-03-10T05:52:39.621470Z","created":"2026-03-10T05:52:39.621470Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T05:53:18.799 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T05:53:18.885 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:18 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/799392' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:19.970 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T05:53:19.970 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph mon dump -f json 2026-03-10T05:53:20.126 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:20.158 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:20.416 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T05:53:20.416 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"100dc3f2-1c45-11f1-8010-b96ca19c2df4","modified":"2026-03-10T05:52:39.621470Z","created":"2026-03-10T05:52:39.621470Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T05:53:20.417 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T05:53:20.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:20 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:20.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:20 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:20.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:20 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:20.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:20 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:20.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:20 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:20.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:20 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:20.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:20 vm01 ceph-mon[47299]: Regenerating cephadm self-signed grafana TLS certificates 2026-03-10T05:53:20.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:20 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:20.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:20 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:20.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:20 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix": "dashboard set-grafana-api-ssl-verify", "value": "false"}]: dispatch 2026-03-10T05:53:20.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:20 vm01 ceph-mon[47299]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-ssl-verify", "value": "false"}]: dispatch 2026-03-10T05:53:20.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:20 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:20.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:20 vm01 ceph-mon[47299]: Deploying daemon grafana.vm01 on vm01 2026-03-10T05:53:20.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:20 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/2485261726' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:21.597 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T05:53:21.597 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph mon dump -f json 2026-03-10T05:53:21.761 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:21.795 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:21.947 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:21 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:22.082 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T05:53:22.082 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"100dc3f2-1c45-11f1-8010-b96ca19c2df4","modified":"2026-03-10T05:52:39.621470Z","created":"2026-03-10T05:52:39.621470Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T05:53:22.082 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T05:53:23.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:22 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/2767968866' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:23.276 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T05:53:23.277 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph mon dump -f json 2026-03-10T05:53:23.441 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:23.481 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:23.737 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T05:53:23.738 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"100dc3f2-1c45-11f1-8010-b96ca19c2df4","modified":"2026-03-10T05:52:39.621470Z","created":"2026-03-10T05:52:39.621470Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T05:53:23.738 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T05:53:24.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:23 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/1658459901' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:24.925 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T05:53:24.926 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph mon dump -f json 2026-03-10T05:53:25.089 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:25.125 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:25.387 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T05:53:25.387 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"100dc3f2-1c45-11f1-8010-b96ca19c2df4","modified":"2026-03-10T05:52:39.621470Z","created":"2026-03-10T05:52:39.621470Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T05:53:25.387 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T05:53:25.453 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:25 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/2955135087' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:26.563 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T05:53:26.563 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph mon dump -f json 2026-03-10T05:53:26.718 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:26.751 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:26.999 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T05:53:26.999 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"100dc3f2-1c45-11f1-8010-b96ca19c2df4","modified":"2026-03-10T05:52:39.621470Z","created":"2026-03-10T05:52:39.621470Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T05:53:26.999 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T05:53:27.330 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:27 vm01 ceph-mon[47299]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:53:27.330 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:27 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:27.330 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:27 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:27.330 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:27 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:27.330 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:27 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:27.330 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:27 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:27.330 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:27 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:27.330 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:27 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:27.330 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:27 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:27.330 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:27 vm01 ceph-mon[47299]: Deploying daemon prometheus.vm01 on vm01 2026-03-10T05:53:27.330 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:27 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/2451375284' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:28.197 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T05:53:28.197 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph mon dump -f json 2026-03-10T05:53:28.362 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:28.398 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:28.661 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T05:53:28.661 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"100dc3f2-1c45-11f1-8010-b96ca19c2df4","modified":"2026-03-10T05:52:39.621470Z","created":"2026-03-10T05:52:39.621470Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T05:53:28.661 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T05:53:29.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:29 vm01 ceph-mon[47299]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:53:29.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:29 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/4020692987' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:29.828 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T05:53:29.829 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph mon dump -f json 2026-03-10T05:53:29.984 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:30.015 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:30.313 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T05:53:30.313 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"100dc3f2-1c45-11f1-8010-b96ca19c2df4","modified":"2026-03-10T05:52:39.621470Z","created":"2026-03-10T05:52:39.621470Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T05:53:30.313 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T05:53:31.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:31 vm01 ceph-mon[47299]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:53:31.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:31 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/45850536' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:31.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:31 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:31.460 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T05:53:31.460 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph mon dump -f json 2026-03-10T05:53:31.626 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:31.654 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:31.901 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T05:53:31.902 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"100dc3f2-1c45-11f1-8010-b96ca19c2df4","modified":"2026-03-10T05:52:39.621470Z","created":"2026-03-10T05:52:39.621470Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T05:53:31.902 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T05:53:32.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:32 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:32.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:32 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:32.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:32 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:32.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:32 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix": "mgr module enable", "module": "prometheus"}]: dispatch 2026-03-10T05:53:32.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:32 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/1988715696' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:33.069 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T05:53:33.069 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph mon dump -f json 2026-03-10T05:53:33.225 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:33.257 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:33.530 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T05:53:33.531 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"100dc3f2-1c45-11f1-8010-b96ca19c2df4","modified":"2026-03-10T05:52:39.621470Z","created":"2026-03-10T05:52:39.621470Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T05:53:33.531 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T05:53:33.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:33 vm01 ceph-mon[47299]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd='[{"prefix": "mgr module enable", "module": "prometheus"}]': finished 2026-03-10T05:53:33.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:33 vm01 ceph-mon[47299]: mgrmap e14: vm01.wforik(active, since 26s) 2026-03-10T05:53:33.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:33 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/2140159908' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:34.674 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T05:53:34.674 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph mon dump -f json 2026-03-10T05:53:34.827 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:34.860 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:35.106 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T05:53:35.106 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"100dc3f2-1c45-11f1-8010-b96ca19c2df4","modified":"2026-03-10T05:52:39.621470Z","created":"2026-03-10T05:52:39.621470Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T05:53:35.107 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T05:53:35.453 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:35 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/1005691130' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:36.314 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T05:53:36.314 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph mon dump -f json 2026-03-10T05:53:36.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:36 vm01 ceph-mon[47299]: Active manager daemon vm01.wforik restarted 2026-03-10T05:53:36.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:36 vm01 ceph-mon[47299]: Activating manager daemon vm01.wforik 2026-03-10T05:53:36.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:36 vm01 ceph-mon[47299]: osdmap e5: 0 total, 0 up, 0 in 2026-03-10T05:53:36.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:36 vm01 ceph-mon[47299]: mgrmap e15: vm01.wforik(active, starting, since 0.235783s) 2026-03-10T05:53:36.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:36 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-10T05:53:36.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:36 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mgr metadata", "who": "vm01.wforik", "id": "vm01.wforik"}]: dispatch 2026-03-10T05:53:36.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:36 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T05:53:36.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:36 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T05:53:36.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:36 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T05:53:36.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:36 vm01 ceph-mon[47299]: Manager daemon vm01.wforik is now available 2026-03-10T05:53:36.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:36 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:36.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:36 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:53:36.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:36 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T05:53:36.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:36 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.wforik/mirror_snapshot_schedule"}]: dispatch 2026-03-10T05:53:36.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:36 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.wforik/trash_purge_schedule"}]: dispatch 2026-03-10T05:53:36.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:36 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:36.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:36 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:36.509 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:36.555 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:36.857 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T05:53:36.857 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"100dc3f2-1c45-11f1-8010-b96ca19c2df4","modified":"2026-03-10T05:52:39.621470Z","created":"2026-03-10T05:52:39.621470Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T05:53:36.857 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T05:53:37.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:37 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:37.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:37 vm01 ceph-mon[47299]: mgrmap e16: vm01.wforik(active, since 1.23982s) 2026-03-10T05:53:37.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:37 vm01 ceph-mon[47299]: [10/Mar/2026:05:53:36] ENGINE Bus STARTING 2026-03-10T05:53:37.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:37 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/2566105202' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:37.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:37 vm01 ceph-mon[47299]: [10/Mar/2026:05:53:36] ENGINE Serving on https://192.168.123.101:7150 2026-03-10T05:53:37.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:37 vm01 ceph-mon[47299]: [10/Mar/2026:05:53:36] ENGINE Client ('192.168.123.101', 57760) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T05:53:37.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:37 vm01 ceph-mon[47299]: [10/Mar/2026:05:53:36] ENGINE Serving on http://192.168.123.101:8765 2026-03-10T05:53:37.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:37 vm01 ceph-mon[47299]: [10/Mar/2026:05:53:36] ENGINE Bus STARTED 2026-03-10T05:53:37.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:37 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:37.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:37 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:37.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:37 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:37.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:37 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:37.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:37 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:53:38.014 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T05:53:38.014 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph mon dump -f json 2026-03-10T05:53:38.209 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:38.255 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:53:38.535 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T05:53:38.535 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"100dc3f2-1c45-11f1-8010-b96ca19c2df4","modified":"2026-03-10T05:52:39.621470Z","created":"2026-03-10T05:52:39.621470Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T05:53:38.536 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T05:53:39.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:39 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:39.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:39 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:39.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:39 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:39.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:39 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:39.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:39 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:53:39.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:39 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:39.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:39 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:53:39.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:39 vm01 ceph-mon[47299]: Updating vm01:/etc/ceph/ceph.conf 2026-03-10T05:53:39.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:39 vm01 ceph-mon[47299]: Updating vm09:/etc/ceph/ceph.conf 2026-03-10T05:53:39.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:39 vm01 ceph-mon[47299]: mgrmap e17: vm01.wforik(active, since 2s) 2026-03-10T05:53:39.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:39 vm01 ceph-mon[47299]: Updating vm01:/var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/config/ceph.conf 2026-03-10T05:53:39.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:39 vm01 ceph-mon[47299]: Updating vm09:/var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/config/ceph.conf 2026-03-10T05:53:39.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:39 vm01 ceph-mon[47299]: Updating vm01:/etc/ceph/ceph.client.admin.keyring 2026-03-10T05:53:39.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:39 vm01 ceph-mon[47299]: Updating vm09:/etc/ceph/ceph.client.admin.keyring 2026-03-10T05:53:39.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:39 vm01 ceph-mon[47299]: Updating vm01:/var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/config/ceph.client.admin.keyring 2026-03-10T05:53:39.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:39 vm01 ceph-mon[47299]: Updating vm09:/var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/config/ceph.client.admin.keyring 2026-03-10T05:53:39.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:39 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/1433152063' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:39.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:39 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:39.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:39 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:39.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:39 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:39.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:39 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:39.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:39 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:39.461 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:39 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm09", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-10T05:53:39.461 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:39 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd='[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm09", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]': finished 2026-03-10T05:53:39.461 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:39 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:39.686 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T05:53:39.686 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph mon dump -f json 2026-03-10T05:53:39.956 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/config/ceph.conf 2026-03-10T05:53:40.224 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T05:53:40.224 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"100dc3f2-1c45-11f1-8010-b96ca19c2df4","modified":"2026-03-10T05:52:39.621470Z","created":"2026-03-10T05:52:39.621470Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T05:53:40.226 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T05:53:40.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:40 vm01 ceph-mon[47299]: Deploying daemon ceph-exporter.vm09 on vm09 2026-03-10T05:53:40.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:40 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:40.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:40 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:40.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:40 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:40.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:40 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:40.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:40 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm09", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-10T05:53:40.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:40 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd='[{"prefix": "auth get-or-create", "entity": "client.crash.vm09", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]': finished 2026-03-10T05:53:40.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:40 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:41.382 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T05:53:41.383 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph mon dump -f json 2026-03-10T05:53:41.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:41 vm01 ceph-mon[47299]: Deploying daemon crash.vm09 on vm09 2026-03-10T05:53:41.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:41 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/3802212979' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:41.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:41 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:41.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:41 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:41.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:41 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:41.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:41 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:41.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:41 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:41.562 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/config/ceph.conf 2026-03-10T05:53:41.835 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T05:53:41.835 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"100dc3f2-1c45-11f1-8010-b96ca19c2df4","modified":"2026-03-10T05:52:39.621470Z","created":"2026-03-10T05:52:39.621470Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T05:53:41.835 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T05:53:42.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:42 vm01 ceph-mon[47299]: Deploying daemon node-exporter.vm09 on vm09 2026-03-10T05:53:42.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:42 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/2045056510' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:43.006 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T05:53:43.007 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph mon dump -f json 2026-03-10T05:53:43.196 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/config/ceph.conf 2026-03-10T05:53:43.523 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T05:53:43.523 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"100dc3f2-1c45-11f1-8010-b96ca19c2df4","modified":"2026-03-10T05:52:39.621470Z","created":"2026-03-10T05:52:39.621470Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T05:53:43.523 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T05:53:43.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:43 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/3481402956' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:44.698 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T05:53:44.699 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph mon dump -f json 2026-03-10T05:53:44.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:44 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:44.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:44 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:44.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:44 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:44.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:44 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:44.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:44 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm09.uxbryv", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T05:53:44.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:44 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.vm09.uxbryv", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-10T05:53:44.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:44 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T05:53:44.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:44 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:44.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:44 vm01 ceph-mon[47299]: Deploying daemon mgr.vm09.uxbryv on vm09 2026-03-10T05:53:44.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:44 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:44.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:44 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:44.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:44 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:44.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:44 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:44.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:44 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T05:53:44.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:44 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:45.152 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/config/ceph.conf 2026-03-10T05:53:45.560 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T05:53:45.560 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"100dc3f2-1c45-11f1-8010-b96ca19c2df4","modified":"2026-03-10T05:52:39.621470Z","created":"2026-03-10T05:52:39.621470Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T05:53:45.560 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T05:53:45.816 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:45 vm09 systemd[1]: Starting Ceph mon.vm09 for 100dc3f2-1c45-11f1-8010-b96ca19c2df4... 2026-03-10T05:53:45.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:45 vm01 ceph-mon[47299]: Deploying daemon mon.vm09 on vm09 2026-03-10T05:53:45.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:45 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/892648051' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:46.061 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:45 vm09 podman[53673]: 2026-03-10 05:53:45.918624346 +0000 UTC m=+0.022341686 container create 98ffb94408d710f491a2c6e9425dccefa04cf62a3cb351de90bda5d4d12a35be (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4-mon-vm09, org.label-schema.name=CentOS Stream 9 Base Image, OSD_FLAVOR=default, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.license=GPLv2, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, org.label-schema.build-date=20260223, CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True, io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9) 2026-03-10T05:53:46.061 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:45 vm09 podman[53673]: 2026-03-10 05:53:45.985304155 +0000 UTC m=+0.089021495 container init 98ffb94408d710f491a2c6e9425dccefa04cf62a3cb351de90bda5d4d12a35be (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4-mon-vm09, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_REF=squid, OSD_FLAVOR=default, ceph=True, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223, FROM_IMAGE=quay.io/centos/centos:stream9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/) 2026-03-10T05:53:46.061 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:45 vm09 podman[53673]: 2026-03-10 05:53:45.989910234 +0000 UTC m=+0.093627574 container start 98ffb94408d710f491a2c6e9425dccefa04cf62a3cb351de90bda5d4d12a35be (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4-mon-vm09, OSD_FLAVOR=default, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, org.label-schema.vendor=CentOS, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.build-date=20260223, org.opencontainers.image.authors=Ceph Release Team , CEPH_REF=squid) 2026-03-10T05:53:46.061 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:45 vm09 bash[53673]: 98ffb94408d710f491a2c6e9425dccefa04cf62a3cb351de90bda5d4d12a35be 2026-03-10T05:53:46.061 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:45 vm09 podman[53673]: 2026-03-10 05:53:45.90685319 +0000 UTC m=+0.010570520 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-10T05:53:46.061 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:45 vm09 systemd[1]: Started Ceph mon.vm09 for 100dc3f2-1c45-11f1-8010-b96ca19c2df4. 2026-03-10T05:53:46.061 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: set uid:gid to 167:167 (ceph:ceph) 2026-03-10T05:53:46.061 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable), process ceph-mon, pid 6 2026-03-10T05:53:46.061 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: pidfile_write: ignore empty --pid-file 2026-03-10T05:53:46.061 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: load: jerasure load: lrc 2026-03-10T05:53:46.061 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: RocksDB version: 7.9.2 2026-03-10T05:53:46.061 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Git sha 0 2026-03-10T05:53:46.061 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Compile date 2026-02-25 18:11:04 2026-03-10T05:53:46.061 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: DB SUMMARY 2026-03-10T05:53:46.061 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: DB Session ID: RU8PX59BXJREPYC42IKR 2026-03-10T05:53:46.061 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: CURRENT file: CURRENT 2026-03-10T05:53:46.061 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: IDENTITY file: IDENTITY 2026-03-10T05:53:46.061 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: MANIFEST file: MANIFEST-000005 size: 59 Bytes 2026-03-10T05:53:46.061 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: SST files in /var/lib/ceph/mon/ceph-vm09/store.db dir, Total Num: 0, files: 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-vm09/store.db: 000004.log size: 511 ; 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.error_if_exists: 0 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.create_if_missing: 0 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.paranoid_checks: 1 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.flush_verify_memtable_count: 1 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.env: 0x55dd2ef92dc0 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.fs: PosixFileSystem 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.info_log: 0x55dd2fa0eb20 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.max_file_opening_threads: 16 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.statistics: (nil) 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.use_fsync: 0 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.max_log_file_size: 0 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.log_file_time_to_roll: 0 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.keep_log_file_num: 1000 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.recycle_log_file_num: 0 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.allow_fallocate: 1 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.allow_mmap_reads: 0 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.allow_mmap_writes: 0 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.use_direct_reads: 0 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.create_missing_column_families: 0 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.db_log_dir: 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.wal_dir: 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.table_cache_numshardbits: 6 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.WAL_ttl_seconds: 0 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.WAL_size_limit_MB: 0 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.is_fd_close_on_exec: 1 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.advise_random_on_open: 1 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.db_write_buffer_size: 0 2026-03-10T05:53:46.062 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.write_buffer_manager: 0x55dd2fa13900 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.use_adaptive_mutex: 0 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.rate_limiter: (nil) 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.wal_recovery_mode: 2 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.enable_thread_tracking: 0 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.enable_pipelined_write: 0 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.unordered_write: 0 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.row_cache: None 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.wal_filter: None 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.allow_ingest_behind: 0 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.two_write_queues: 0 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.manual_wal_flush: 0 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.wal_compression: 0 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.atomic_flush: 0 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.persist_stats_to_disk: 0 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.write_dbid_to_manifest: 0 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.log_readahead_size: 0 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.best_efforts_recovery: 0 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.allow_data_in_errors: 0 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.db_host_id: __hostname__ 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.enforce_single_del_contracts: true 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.max_background_jobs: 2 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.max_background_compactions: -1 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.max_subcompactions: 1 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.delayed_write_rate : 16777216 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.max_total_wal_size: 0 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.stats_dump_period_sec: 600 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.stats_persist_period_sec: 600 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.max_open_files: -1 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.bytes_per_sync: 0 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.wal_bytes_per_sync: 0 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.strict_bytes_per_sync: 0 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.compaction_readahead_size: 0 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.max_background_flushes: -1 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Compression algorithms supported: 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: kZSTD supported: 0 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: kXpressCompression supported: 0 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: kBZip2Compression supported: 0 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: kLZ4Compression supported: 1 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: kZlibCompression supported: 1 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: kLZ4HCCompression supported: 1 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: kSnappyCompression supported: 1 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Fast CRC32 supported: Supported on x86 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: DMutex implementation: pthread_mutex_t 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-vm09/store.db/MANIFEST-000005 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-10T05:53:46.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.merge_operator: 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.compaction_filter: None 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.compaction_filter_factory: None 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.sst_partitioner_factory: None 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.memtable_factory: SkipListFactory 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.table_factory: BlockBasedTable 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55dd2fa0e6e0) 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: cache_index_and_filter_blocks: 1 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: pin_top_level_index_and_filter: 1 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: index_type: 0 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: data_block_index_type: 0 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: index_shortening: 1 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: data_block_hash_table_util_ratio: 0.750000 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: checksum: 4 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: no_block_cache: 0 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: block_cache: 0x55dd2fa33350 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: block_cache_name: BinnedLRUCache 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: block_cache_options: 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: capacity : 536870912 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: num_shard_bits : 4 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: strict_capacity_limit : 0 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: high_pri_pool_ratio: 0.000 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: block_cache_compressed: (nil) 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: persistent_cache: (nil) 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: block_size: 4096 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: block_size_deviation: 10 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: block_restart_interval: 16 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: index_block_restart_interval: 1 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: metadata_block_size: 4096 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: partition_filters: 0 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: use_delta_encoding: 1 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: filter_policy: bloomfilter 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: whole_key_filtering: 1 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: verify_compression: 0 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: read_amp_bytes_per_bit: 0 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: format_version: 5 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: enable_index_compression: 1 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: block_align: 0 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: max_auto_readahead_size: 262144 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: prepopulate_block_cache: 0 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: initial_auto_readahead_size: 8192 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout: num_file_reads_for_auto_readahead: 2 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.write_buffer_size: 33554432 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.max_write_buffer_number: 2 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.compression: NoCompression 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.bottommost_compression: Disabled 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.prefix_extractor: nullptr 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.num_levels: 7 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.compression_opts.window_bits: -14 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.compression_opts.level: 32767 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.compression_opts.strategy: 0 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-10T05:53:46.064 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.compression_opts.enabled: false 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.target_file_size_base: 67108864 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.target_file_size_multiplier: 1 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.arena_block_size: 1048576 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.disable_auto_compactions: 0 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.inplace_update_support: 0 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.inplace_update_num_locks: 10000 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.memtable_huge_page_size: 0 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.bloom_locality: 0 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.max_successive_merges: 0 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.optimize_filters_for_hits: 0 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.paranoid_file_checks: 0 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.force_consistency_checks: 1 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.report_bg_io_stats: 0 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.ttl: 2592000 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.periodic_compaction_seconds: 0 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.enable_blob_files: false 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.min_blob_size: 0 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.blob_file_size: 268435456 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.blob_compression_type: NoCompression 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.enable_blob_garbage_collection: false 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.blob_file_starting_level: 0 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-vm09/store.db/MANIFEST-000005 succeeded,manifest_file_number is 5, next_file_number is 7, last_sequence is 0, log_number is 0,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 0 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 0 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 00bd99f4-2aa3-45c4-b713-c632ac93a454 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773122026020137, "job": 1, "event": "recovery_started", "wal_files": [4]} 2026-03-10T05:53:46.065 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #4 mode 2 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773122026020737, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 8, "file_size": 1643, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 1, "largest_seqno": 5, "table_properties": {"data_size": 523, "index_size": 31, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 69, "raw_key_size": 115, "raw_average_key_size": 23, "raw_value_size": 401, "raw_average_value_size": 80, "num_data_blocks": 1, "num_entries": 5, "num_filter_entries": 5, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1773122026, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "00bd99f4-2aa3-45c4-b713-c632ac93a454", "db_session_id": "RU8PX59BXJREPYC42IKR", "orig_file_number": 8, "seqno_to_time_mapping": "N/A"}} 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773122026020795, "job": 1, "event": "recovery_finished"} 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: [db/version_set.cc:5047] Creating manifest 10 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-vm09/store.db/000004.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55dd2fa34e00 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: DB pointer 0x55dd2fb4e000 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mon.vm09 does not exist in monmap, will attempt to join an existing cluster 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: ** DB Stats ** 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: ** Compaction Stats [default] ** 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: L0 1/0 1.60 KB 0.2 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 2.7 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: Sum 1/0 1.60 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 2.7 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 2.7 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: ** Compaction Stats [default] ** 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 2.7 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: Flush(GB): cumulative 0.000, interval 0.000 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: AddFile(Total Files): cumulative 0, interval 0 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: AddFile(L0 Files): cumulative 0, interval 0 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: AddFile(Keys): cumulative 0, interval 0 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: Cumulative compaction: 0.00 GB write, 0.12 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: Interval compaction: 0.00 GB write, 0.12 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: Block cache BinnedLRUCache@0x55dd2fa33350#6 capacity: 512.00 MB usage: 0.86 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 8e-06 secs_since: 0 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: Block cache entry stats(count,size,portion): DataBlock(1,0.64 KB,0.00012219%) FilterBlock(1,0.11 KB,2.08616e-05%) IndexBlock(1,0.11 KB,2.08616e-05%) Misc(1,0.00 KB,0%) 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: ** File Read Latency Histogram By Level [default] ** 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: using public_addr v2:192.168.123.109:0/0 -> [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: starting mon.vm09 rank -1 at public addrs [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] at bind addrs [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon_data /var/lib/ceph/mon/ceph-vm09 fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mon.vm09@-1(???) e0 preinit fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mon.vm09@-1(synchronizing).mds e1 new map 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mon.vm09@-1(synchronizing).mds e1 print_map 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: e1 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: btime 2026-03-10T05:52:41:179707+0000 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: enable_multiple, ever_enabled_multiple: 1,1 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: legacy client fscid: -1 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout: No filesystems configured 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mon.vm09@-1(synchronizing).osd e0 _set_cache_ratios kv ratio 0.25 inc ratio 0.375 full ratio 0.375 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mon.vm09@-1(synchronizing).osd e0 register_cache_with_pcm pcm target: 2147483648 pcm max: 1020054732 pcm min: 134217728 inc_osd_cache size: 1 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mon.vm09@-1(synchronizing).osd e1 e1: 0 total, 0 up, 0 in 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mon.vm09@-1(synchronizing).osd e2 e2: 0 total, 0 up, 0 in 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mon.vm09@-1(synchronizing).osd e3 e3: 0 total, 0 up, 0 in 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mon.vm09@-1(synchronizing).osd e4 e4: 0 total, 0 up, 0 in 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mon.vm09@-1(synchronizing).osd e5 e5: 0 total, 0 up, 0 in 2026-03-10T05:53:46.066 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mon.vm09@-1(synchronizing).osd e5 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mon.vm09@-1(synchronizing).osd e5 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mon.vm09@-1(synchronizing).osd e5 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mon.vm09@-1(synchronizing).osd e5 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mkfs 100dc3f2-1c45-11f1-8010-b96ca19c2df4 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mon.vm01 is new leader, mons vm01 in quorum (ranks 0) 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mon.vm01 is new leader, mons vm01 in quorum (ranks 0) 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: monmap epoch 1 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: last_changed 2026-03-10T05:52:39.621470+0000 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: created 2026-03-10T05:52:39.621470+0000 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: min_mon_release 19 (squid) 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: election_strategy: 1 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.vm01 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: fsmap 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: osdmap e1: 0 total, 0 up, 0 in 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mgrmap e1: no daemons active 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/3186782452' entity='client.admin' 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/2171342023' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/344241115' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Activating manager daemon vm01.wforik 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mgrmap e2: vm01.wforik(active, starting, since 0.0038552s) 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14100 192.168.123.101:0/1126556124' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14100 192.168.123.101:0/1126556124' entity='mgr.vm01.wforik' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14100 192.168.123.101:0/1126556124' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14100 192.168.123.101:0/1126556124' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14100 192.168.123.101:0/1126556124' entity='mgr.vm01.wforik' cmd=[{"prefix": "mgr metadata", "who": "vm01.wforik", "id": "vm01.wforik"}]: dispatch 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Manager daemon vm01.wforik is now available 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14100 192.168.123.101:0/1126556124' entity='mgr.vm01.wforik' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.wforik/mirror_snapshot_schedule"}]: dispatch 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14100 192.168.123.101:0/1126556124' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14100 192.168.123.101:0/1126556124' entity='mgr.vm01.wforik' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.wforik/trash_purge_schedule"}]: dispatch 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14100 192.168.123.101:0/1126556124' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14100 192.168.123.101:0/1126556124' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mgrmap e3: vm01.wforik(active, since 1.83786s) 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/256650801' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/2438874528' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mgrmap e4: vm01.wforik(active, since 2s) 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/1968105725' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/1968105725' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mgrmap e5: vm01.wforik(active, since 3s) 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/3286629993' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Active manager daemon vm01.wforik restarted 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Activating manager daemon vm01.wforik 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: osdmap e2: 0 total, 0 up, 0 in 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mgrmap e6: vm01.wforik(active, starting, since 0.00395s) 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' cmd=[{"prefix": "mgr metadata", "who": "vm01.wforik", "id": "vm01.wforik"}]: dispatch 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Manager daemon vm01.wforik is now available 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.wforik/mirror_snapshot_schedule"}]: dispatch 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.wforik/trash_purge_schedule"}]: dispatch 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Found migration_current of "None". Setting to last migration. 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mgrmap e7: vm01.wforik(active, since 1.00718s) 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: [10/Mar/2026:05:52:54] ENGINE Bus STARTING 2026-03-10T05:53:46.067 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: [10/Mar/2026:05:52:54] ENGINE Serving on http://192.168.123.101:8765 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: [10/Mar/2026:05:52:54] ENGINE Serving on https://192.168.123.101:7150 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: [10/Mar/2026:05:52:54] ENGINE Bus STARTED 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: [10/Mar/2026:05:52:54] ENGINE Client ('192.168.123.101', 60110) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Generating ssh key... 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mgrmap e8: vm01.wforik(active, since 2s) 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm01", "addr": "192.168.123.101", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Deploying cephadm binary to vm01 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Added host vm01 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Saving service mon spec with placement count:5 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Saving service mgr spec with placement count:2 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.14144 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "crash", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Saving service crash spec with placement * 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.14146 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "ceph-exporter", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Saving service ceph-exporter spec with placement * 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.14148 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "prometheus", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Saving service prometheus spec with placement count:1 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.14150 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "grafana", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Saving service grafana spec with placement count:1 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14118 192.168.123.101:0/4203552096' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.14152 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "node-exporter", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Saving service node-exporter spec with placement * 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.14154 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "alertmanager", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Saving service alertmanager spec with placement count:1 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/2380699250' entity='client.admin' 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/3486127958' entity='client.admin' 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/1832610603' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/1832610603' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mgrmap e9: vm01.wforik(active, since 9s) 2026-03-10T05:53:46.068 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/4273479382' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Active manager daemon vm01.wforik restarted 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Activating manager daemon vm01.wforik 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: osdmap e3: 0 total, 0 up, 0 in 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mgrmap e10: vm01.wforik(active, starting, since 0.00526092s) 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix": "mgr metadata", "who": "vm01.wforik", "id": "vm01.wforik"}]: dispatch 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Manager daemon vm01.wforik is now available 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.wforik/mirror_snapshot_schedule"}]: dispatch 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.wforik/trash_purge_schedule"}]: dispatch 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: [10/Mar/2026:05:53:06] ENGINE Bus STARTING 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mgrmap e11: vm01.wforik(active, since 1.00872s) 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.14166 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.14166 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: [10/Mar/2026:05:53:06] ENGINE Serving on https://192.168.123.101:7150 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: [10/Mar/2026:05:53:06] ENGINE Client ('192.168.123.101', 54832) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: [10/Mar/2026:05:53:06] ENGINE Serving on http://192.168.123.101:8765 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: [10/Mar/2026:05:53:06] ENGINE Bus STARTED 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.14174 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.14176 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/1548582451' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/3276924622' entity='client.admin' 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mgrmap e12: vm01.wforik(active, since 2s) 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/278616664' entity='client.admin' 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm01", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd='[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm01", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]': finished 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Deploying daemon ceph-exporter.vm01 on vm01 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.14184 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.14186 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm09", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Deploying cephadm binary to vm09 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mgrmap e13: vm01.wforik(active, since 6s) 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm01", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd='[{"prefix": "auth get-or-create", "entity": "client.crash.vm01", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]': finished 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Deploying daemon crash.vm01 on vm01 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Added host vm09 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Deploying daemon node-exporter.vm01 on vm01 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.14189 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/3876165708' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/3876165708' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.14193 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "2;vm01:192.168.123.101=vm01;vm09:192.168.123.109=vm09", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:46.069 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Saving service mon spec with placement vm01:192.168.123.101=vm01;vm09:192.168.123.109=vm09;count:2 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/285557039' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Deploying daemon alertmanager.vm01 on vm01 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/951122430' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/799392' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Regenerating cephadm self-signed grafana TLS certificates 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix": "dashboard set-grafana-api-ssl-verify", "value": "false"}]: dispatch 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-ssl-verify", "value": "false"}]: dispatch 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Deploying daemon grafana.vm01 on vm01 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/2485261726' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/2767968866' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/1658459901' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/2955135087' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Deploying daemon prometheus.vm01 on vm01 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/2451375284' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/4020692987' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/45850536' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd=[{"prefix": "mgr module enable", "module": "prometheus"}]: dispatch 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/1988715696' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14162 192.168.123.101:0/935748757' entity='mgr.vm01.wforik' cmd='[{"prefix": "mgr module enable", "module": "prometheus"}]': finished 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mgrmap e14: vm01.wforik(active, since 26s) 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/2140159908' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:46.070 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/1005691130' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Active manager daemon vm01.wforik restarted 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Activating manager daemon vm01.wforik 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: osdmap e5: 0 total, 0 up, 0 in 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mgrmap e15: vm01.wforik(active, starting, since 0.235783s) 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mgr metadata", "who": "vm01.wforik", "id": "vm01.wforik"}]: dispatch 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Manager daemon vm01.wforik is now available 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.wforik/mirror_snapshot_schedule"}]: dispatch 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.wforik/trash_purge_schedule"}]: dispatch 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mgrmap e16: vm01.wforik(active, since 1.23982s) 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: [10/Mar/2026:05:53:36] ENGINE Bus STARTING 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/2566105202' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: [10/Mar/2026:05:53:36] ENGINE Serving on https://192.168.123.101:7150 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: [10/Mar/2026:05:53:36] ENGINE Client ('192.168.123.101', 57760) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: [10/Mar/2026:05:53:36] ENGINE Serving on http://192.168.123.101:8765 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: [10/Mar/2026:05:53:36] ENGINE Bus STARTED 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Updating vm01:/etc/ceph/ceph.conf 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Updating vm09:/etc/ceph/ceph.conf 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mgrmap e17: vm01.wforik(active, since 2s) 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Updating vm01:/var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/config/ceph.conf 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Updating vm09:/var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/config/ceph.conf 2026-03-10T05:53:46.367 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Updating vm01:/etc/ceph/ceph.client.admin.keyring 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Updating vm09:/etc/ceph/ceph.client.admin.keyring 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Updating vm01:/var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/config/ceph.client.admin.keyring 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Updating vm09:/var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/config/ceph.client.admin.keyring 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/1433152063' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm09", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd='[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm09", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]': finished 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Deploying daemon ceph-exporter.vm09 on vm09 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm09", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd='[{"prefix": "auth get-or-create", "entity": "client.crash.vm09", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]': finished 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Deploying daemon crash.vm09 on vm09 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/3802212979' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Deploying daemon node-exporter.vm09 on vm09 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/2045056510' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/3481402956' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm09.uxbryv", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.vm09.uxbryv", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Deploying daemon mgr.vm09.uxbryv on vm09 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: Deploying daemon mon.vm09 on vm09 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/892648051' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:46.368 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:46 vm09 ceph-mon[53707]: mon.vm09@-1(synchronizing).paxosservice(auth 1..8) refresh upgraded, format 0 -> 3 2026-03-10T05:53:46.864 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T05:53:46.864 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph mon dump -f json 2026-03-10T05:53:47.049 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm09/config 2026-03-10T05:53:51.419 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-10T05:53:51.419 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T05:53:51.419 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: mon.vm01 calling monitor election 2026-03-10T05:53:51.419 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:53:51.420 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T05:53:51.420 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: from='mgr.? 192.168.123.109:0/3249151571' entity='mgr.vm09.uxbryv' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm09.uxbryv/crt"}]: dispatch 2026-03-10T05:53:51.420 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T05:53:51.420 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: mon.vm09 calling monitor election 2026-03-10T05:53:51.420 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T05:53:51.420 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T05:53:51.420 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T05:53:51.420 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T05:53:51.420 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: mon.vm01 is new leader, mons vm01,vm09 in quorum (ranks 0,1) 2026-03-10T05:53:51.420 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: monmap epoch 2 2026-03-10T05:53:51.420 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 2026-03-10T05:53:51.420 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: last_changed 2026-03-10T05:53:46.079053+0000 2026-03-10T05:53:51.420 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: created 2026-03-10T05:52:39.621470+0000 2026-03-10T05:53:51.420 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: min_mon_release 19 (squid) 2026-03-10T05:53:51.420 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: election_strategy: 1 2026-03-10T05:53:51.420 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.vm01 2026-03-10T05:53:51.420 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.vm09 2026-03-10T05:53:51.420 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: fsmap 2026-03-10T05:53:51.420 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: osdmap e5: 0 total, 0 up, 0 in 2026-03-10T05:53:51.420 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: mgrmap e17: vm01.wforik(active, since 15s) 2026-03-10T05:53:51.420 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: overall HEALTH_OK 2026-03-10T05:53:51.420 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: Standby manager daemon vm09.uxbryv started 2026-03-10T05:53:51.420 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:51.420 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: from='mgr.? 192.168.123.109:0/3249151571' entity='mgr.vm09.uxbryv' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-10T05:53:51.420 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: from='mgr.? 192.168.123.109:0/3249151571' entity='mgr.vm09.uxbryv' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm09.uxbryv/key"}]: dispatch 2026-03-10T05:53:51.420 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:51 vm09 ceph-mon[53707]: from='mgr.? 192.168.123.109:0/3249151571' entity='mgr.vm09.uxbryv' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: mon.vm01 calling monitor election 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: from='mgr.? 192.168.123.109:0/3249151571' entity='mgr.vm09.uxbryv' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm09.uxbryv/crt"}]: dispatch 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: mon.vm09 calling monitor election 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: mon.vm01 is new leader, mons vm01,vm09 in quorum (ranks 0,1) 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: monmap epoch 2 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: last_changed 2026-03-10T05:53:46.079053+0000 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: created 2026-03-10T05:52:39.621470+0000 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: min_mon_release 19 (squid) 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: election_strategy: 1 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.vm01 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.vm09 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: fsmap 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: osdmap e5: 0 total, 0 up, 0 in 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: mgrmap e17: vm01.wforik(active, since 15s) 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: overall HEALTH_OK 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: Standby manager daemon vm09.uxbryv started 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: from='mgr.? 192.168.123.109:0/3249151571' entity='mgr.vm09.uxbryv' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: from='mgr.? 192.168.123.109:0/3249151571' entity='mgr.vm09.uxbryv' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm09.uxbryv/key"}]: dispatch 2026-03-10T05:53:51.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:51 vm01 ceph-mon[47299]: from='mgr.? 192.168.123.109:0/3249151571' entity='mgr.vm09.uxbryv' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-10T05:53:51.741 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T05:53:51.741 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":2,"fsid":"100dc3f2-1c45-11f1-8010-b96ca19c2df4","modified":"2026-03-10T05:53:46.079053Z","created":"2026-03-10T05:52:39.621470Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"vm09","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:3300","nonce":0},{"type":"v1","addr":"192.168.123.109:6789","nonce":0}]},"addr":"192.168.123.109:6789/0","public_addr":"192.168.123.109:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1]} 2026-03-10T05:53:51.741 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 2 2026-03-10T05:53:51.902 INFO:tasks.cephadm:Generating final ceph.conf file... 2026-03-10T05:53:51.902 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph config generate-minimal-conf 2026-03-10T05:53:52.125 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:53:52.178 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:52 vm09 ceph-mon[53707]: mgrmap e18: vm01.wforik(active, since 15s), standbys: vm09.uxbryv 2026-03-10T05:53:52.178 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:52 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mgr metadata", "who": "vm09.uxbryv", "id": "vm09.uxbryv"}]: dispatch 2026-03-10T05:53:52.188 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:52 vm01 ceph-mon[47299]: mgrmap e18: vm01.wforik(active, since 15s), standbys: vm09.uxbryv 2026-03-10T05:53:52.188 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:52 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mgr metadata", "who": "vm09.uxbryv", "id": "vm09.uxbryv"}]: dispatch 2026-03-10T05:53:52.188 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:52 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:52.188 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:52 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:52.188 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:52 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:52.188 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:52 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:53:52.188 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:52 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/1722381690' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:52.188 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:52 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:52.188 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:52 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:52.188 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:52 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:52.188 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:52 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:52.189 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:52 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:52.189 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:52 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T05:53:52.189 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:52 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T05:53:52.189 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:52 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:52.189 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:52 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T05:53:52.409 INFO:teuthology.orchestra.run.vm01.stdout:# minimal ceph.conf for 100dc3f2-1c45-11f1-8010-b96ca19c2df4 2026-03-10T05:53:52.409 INFO:teuthology.orchestra.run.vm01.stdout:[global] 2026-03-10T05:53:52.409 INFO:teuthology.orchestra.run.vm01.stdout: fsid = 100dc3f2-1c45-11f1-8010-b96ca19c2df4 2026-03-10T05:53:52.409 INFO:teuthology.orchestra.run.vm01.stdout: mon_host = [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] 2026-03-10T05:53:52.459 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:52 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:52.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:52 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:52.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:52 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:52.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:52 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:53:52.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:52 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/1722381690' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:53:52.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:52 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:52.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:52 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:52.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:52 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:52.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:52 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:52.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:52 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:52.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:52 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T05:53:52.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:52 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T05:53:52.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:52 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:52.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:52 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T05:53:52.567 INFO:tasks.cephadm:Distributing (final) config and client.admin keyring... 2026-03-10T05:53:52.567 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T05:53:52.567 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/ceph/ceph.conf 2026-03-10T05:53:52.599 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T05:53:52.599 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T05:53:52.664 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T05:53:52.664 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/ceph/ceph.conf 2026-03-10T05:53:52.691 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T05:53:52.692 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T05:53:52.757 INFO:tasks.cephadm:Deploying OSDs... 2026-03-10T05:53:52.757 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T05:53:52.757 DEBUG:teuthology.orchestra.run.vm01:> dd if=/scratch_devs of=/dev/stdout 2026-03-10T05:53:52.785 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T05:53:52.785 DEBUG:teuthology.orchestra.run.vm01:> ls /dev/[sv]d? 2026-03-10T05:53:52.843 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vda 2026-03-10T05:53:52.843 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vdb 2026-03-10T05:53:52.843 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vdc 2026-03-10T05:53:52.843 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vdd 2026-03-10T05:53:52.843 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vde 2026-03-10T05:53:52.843 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-10T05:53:52.843 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-10T05:53:52.843 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vdb 2026-03-10T05:53:52.903 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vdb 2026-03-10T05:53:52.903 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T05:53:52.903 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 223 Links: 1 Device type: fc,10 2026-03-10T05:53:52.903 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T05:53:52.903 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T05:53:52.903 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-10 05:53:10.820474698 +0000 2026-03-10T05:53:52.904 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-10 05:50:34.953527964 +0000 2026-03-10T05:53:52.904 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-10 05:50:34.953527964 +0000 2026-03-10T05:53:52.904 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-10 05:47:43.309000000 +0000 2026-03-10T05:53:52.904 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-10T05:53:52.974 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-10T05:53:52.974 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-10T05:53:52.974 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.00011779 s, 4.3 MB/s 2026-03-10T05:53:52.975 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-10T05:53:53.034 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vdc 2026-03-10T05:53:53.104 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vdc 2026-03-10T05:53:53.104 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T05:53:53.104 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 250 Links: 1 Device type: fc,20 2026-03-10T05:53:53.104 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T05:53:53.104 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T05:53:53.104 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-10 05:53:10.905474842 +0000 2026-03-10T05:53:53.104 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-10 05:50:35.004528004 +0000 2026-03-10T05:53:53.104 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-10 05:50:35.004528004 +0000 2026-03-10T05:53:53.104 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-10 05:47:43.327000000 +0000 2026-03-10T05:53:53.104 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-10T05:53:53.190 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-10T05:53:53.190 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-10T05:53:53.190 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.00129172 s, 396 kB/s 2026-03-10T05:53:53.192 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-10T05:53:53.221 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vdd 2026-03-10T05:53:53.278 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vdd 2026-03-10T05:53:53.279 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T05:53:53.279 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 253 Links: 1 Device type: fc,30 2026-03-10T05:53:53.279 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T05:53:53.279 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T05:53:53.279 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-10 05:53:10.957474930 +0000 2026-03-10T05:53:53.279 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-10 05:50:35.005528005 +0000 2026-03-10T05:53:53.279 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-10 05:50:35.005528005 +0000 2026-03-10T05:53:53.279 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-10 05:47:43.350000000 +0000 2026-03-10T05:53:53.279 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-10T05:53:53.311 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:53 vm01 ceph-mon[47299]: Updating vm01:/etc/ceph/ceph.conf 2026-03-10T05:53:53.311 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:53 vm01 ceph-mon[47299]: Updating vm09:/etc/ceph/ceph.conf 2026-03-10T05:53:53.311 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:53 vm01 ceph-mon[47299]: Updating vm01:/var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/config/ceph.conf 2026-03-10T05:53:53.311 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:53 vm01 ceph-mon[47299]: Updating vm09:/var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/config/ceph.conf 2026-03-10T05:53:53.311 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:53 vm01 ceph-mon[47299]: Reconfiguring mon.vm01 (unknown last config time)... 2026-03-10T05:53:53.311 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:53 vm01 ceph-mon[47299]: Reconfiguring daemon mon.vm01 on vm01 2026-03-10T05:53:53.311 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:53 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/995490876' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:53.311 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:53 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:53.311 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:53 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:53.311 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:53 vm01 ceph-mon[47299]: Reconfiguring mgr.vm01.wforik (unknown last config time)... 2026-03-10T05:53:53.311 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:53 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm01.wforik", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T05:53:53.311 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:53 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T05:53:53.311 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:53 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:53.311 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:53 vm01 ceph-mon[47299]: Reconfiguring daemon mgr.vm01.wforik on vm01 2026-03-10T05:53:53.311 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:53 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:53.311 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:53 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:53.311 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:53 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm01", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-10T05:53:53.311 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:53 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:53.351 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-10T05:53:53.351 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-10T05:53:53.351 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 8.9278e-05 s, 5.7 MB/s 2026-03-10T05:53:53.355 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-10T05:53:53.421 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vde 2026-03-10T05:53:53.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:53 vm09 ceph-mon[53707]: Updating vm01:/etc/ceph/ceph.conf 2026-03-10T05:53:53.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:53 vm09 ceph-mon[53707]: Updating vm09:/etc/ceph/ceph.conf 2026-03-10T05:53:53.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:53 vm09 ceph-mon[53707]: Updating vm01:/var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/config/ceph.conf 2026-03-10T05:53:53.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:53 vm09 ceph-mon[53707]: Updating vm09:/var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/config/ceph.conf 2026-03-10T05:53:53.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:53 vm09 ceph-mon[53707]: Reconfiguring mon.vm01 (unknown last config time)... 2026-03-10T05:53:53.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:53 vm09 ceph-mon[53707]: Reconfiguring daemon mon.vm01 on vm01 2026-03-10T05:53:53.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:53 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/995490876' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:53.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:53 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:53.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:53 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:53.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:53 vm09 ceph-mon[53707]: Reconfiguring mgr.vm01.wforik (unknown last config time)... 2026-03-10T05:53:53.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:53 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm01.wforik", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T05:53:53.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:53 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T05:53:53.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:53 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:53.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:53 vm09 ceph-mon[53707]: Reconfiguring daemon mgr.vm01.wforik on vm01 2026-03-10T05:53:53.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:53 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:53.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:53 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:53.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:53 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm01", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-10T05:53:53.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:53 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:53.478 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vde 2026-03-10T05:53:53.478 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T05:53:53.478 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-10T05:53:53.478 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T05:53:53.478 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T05:53:53.478 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-10 05:53:11.032475057 +0000 2026-03-10T05:53:53.478 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-10 05:50:35.002528003 +0000 2026-03-10T05:53:53.478 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-10 05:50:35.002528003 +0000 2026-03-10T05:53:53.478 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-10 05:47:43.354000000 +0000 2026-03-10T05:53:53.479 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-10T05:53:53.565 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-10T05:53:53.565 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-10T05:53:53.565 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.000147395 s, 3.5 MB/s 2026-03-10T05:53:53.566 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-10T05:53:53.592 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T05:53:53.592 DEBUG:teuthology.orchestra.run.vm09:> dd if=/scratch_devs of=/dev/stdout 2026-03-10T05:53:53.607 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T05:53:53.607 DEBUG:teuthology.orchestra.run.vm09:> ls /dev/[sv]d? 2026-03-10T05:53:53.662 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vda 2026-03-10T05:53:53.663 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vdb 2026-03-10T05:53:53.663 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vdc 2026-03-10T05:53:53.663 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vdd 2026-03-10T05:53:53.663 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vde 2026-03-10T05:53:53.663 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-10T05:53:53.663 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-10T05:53:53.663 DEBUG:teuthology.orchestra.run.vm09:> stat /dev/vdb 2026-03-10T05:53:53.724 INFO:teuthology.orchestra.run.vm09.stdout: File: /dev/vdb 2026-03-10T05:53:53.724 INFO:teuthology.orchestra.run.vm09.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T05:53:53.724 INFO:teuthology.orchestra.run.vm09.stdout:Device: 6h/6d Inode: 254 Links: 1 Device type: fc,10 2026-03-10T05:53:53.724 INFO:teuthology.orchestra.run.vm09.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T05:53:53.724 INFO:teuthology.orchestra.run.vm09.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T05:53:53.724 INFO:teuthology.orchestra.run.vm09.stdout:Access: 2026-03-10 05:53:37.810215931 +0000 2026-03-10T05:53:53.724 INFO:teuthology.orchestra.run.vm09.stdout:Modify: 2026-03-10 05:50:36.484014067 +0000 2026-03-10T05:53:53.724 INFO:teuthology.orchestra.run.vm09.stdout:Change: 2026-03-10 05:50:36.484014067 +0000 2026-03-10T05:53:53.724 INFO:teuthology.orchestra.run.vm09.stdout: Birth: 2026-03-10 05:48:13.297000000 +0000 2026-03-10T05:53:53.725 DEBUG:teuthology.orchestra.run.vm09:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-10T05:53:53.794 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records in 2026-03-10T05:53:53.794 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records out 2026-03-10T05:53:53.794 INFO:teuthology.orchestra.run.vm09.stderr:512 bytes copied, 0.000113673 s, 4.5 MB/s 2026-03-10T05:53:53.795 DEBUG:teuthology.orchestra.run.vm09:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-10T05:53:53.856 DEBUG:teuthology.orchestra.run.vm09:> stat /dev/vdc 2026-03-10T05:53:53.913 INFO:teuthology.orchestra.run.vm09.stdout: File: /dev/vdc 2026-03-10T05:53:53.914 INFO:teuthology.orchestra.run.vm09.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T05:53:53.914 INFO:teuthology.orchestra.run.vm09.stdout:Device: 6h/6d Inode: 255 Links: 1 Device type: fc,20 2026-03-10T05:53:53.914 INFO:teuthology.orchestra.run.vm09.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T05:53:53.914 INFO:teuthology.orchestra.run.vm09.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T05:53:53.914 INFO:teuthology.orchestra.run.vm09.stdout:Access: 2026-03-10 05:53:37.831215941 +0000 2026-03-10T05:53:53.914 INFO:teuthology.orchestra.run.vm09.stdout:Modify: 2026-03-10 05:50:36.485014068 +0000 2026-03-10T05:53:53.914 INFO:teuthology.orchestra.run.vm09.stdout:Change: 2026-03-10 05:50:36.485014068 +0000 2026-03-10T05:53:53.914 INFO:teuthology.orchestra.run.vm09.stdout: Birth: 2026-03-10 05:48:13.304000000 +0000 2026-03-10T05:53:53.914 DEBUG:teuthology.orchestra.run.vm09:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-10T05:53:53.977 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records in 2026-03-10T05:53:53.977 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records out 2026-03-10T05:53:53.977 INFO:teuthology.orchestra.run.vm09.stderr:512 bytes copied, 0.000111448 s, 4.6 MB/s 2026-03-10T05:53:53.978 DEBUG:teuthology.orchestra.run.vm09:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-10T05:53:54.036 DEBUG:teuthology.orchestra.run.vm09:> stat /dev/vdd 2026-03-10T05:53:54.094 INFO:teuthology.orchestra.run.vm09.stdout: File: /dev/vdd 2026-03-10T05:53:54.094 INFO:teuthology.orchestra.run.vm09.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T05:53:54.094 INFO:teuthology.orchestra.run.vm09.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-10T05:53:54.094 INFO:teuthology.orchestra.run.vm09.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T05:53:54.094 INFO:teuthology.orchestra.run.vm09.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T05:53:54.094 INFO:teuthology.orchestra.run.vm09.stdout:Access: 2026-03-10 05:53:37.851215950 +0000 2026-03-10T05:53:54.094 INFO:teuthology.orchestra.run.vm09.stdout:Modify: 2026-03-10 05:50:36.488014071 +0000 2026-03-10T05:53:54.094 INFO:teuthology.orchestra.run.vm09.stdout:Change: 2026-03-10 05:50:36.488014071 +0000 2026-03-10T05:53:54.094 INFO:teuthology.orchestra.run.vm09.stdout: Birth: 2026-03-10 05:48:13.310000000 +0000 2026-03-10T05:53:54.094 DEBUG:teuthology.orchestra.run.vm09:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-10T05:53:54.159 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records in 2026-03-10T05:53:54.159 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records out 2026-03-10T05:53:54.159 INFO:teuthology.orchestra.run.vm09.stderr:512 bytes copied, 0.000113361 s, 4.5 MB/s 2026-03-10T05:53:54.160 DEBUG:teuthology.orchestra.run.vm09:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-10T05:53:54.217 DEBUG:teuthology.orchestra.run.vm09:> stat /dev/vde 2026-03-10T05:53:54.276 INFO:teuthology.orchestra.run.vm09.stdout: File: /dev/vde 2026-03-10T05:53:54.276 INFO:teuthology.orchestra.run.vm09.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T05:53:54.276 INFO:teuthology.orchestra.run.vm09.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-10T05:53:54.276 INFO:teuthology.orchestra.run.vm09.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T05:53:54.276 INFO:teuthology.orchestra.run.vm09.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T05:53:54.276 INFO:teuthology.orchestra.run.vm09.stdout:Access: 2026-03-10 05:53:37.872215960 +0000 2026-03-10T05:53:54.277 INFO:teuthology.orchestra.run.vm09.stdout:Modify: 2026-03-10 05:50:36.499014082 +0000 2026-03-10T05:53:54.277 INFO:teuthology.orchestra.run.vm09.stdout:Change: 2026-03-10 05:50:36.499014082 +0000 2026-03-10T05:53:54.277 INFO:teuthology.orchestra.run.vm09.stdout: Birth: 2026-03-10 05:48:13.319000000 +0000 2026-03-10T05:53:54.277 DEBUG:teuthology.orchestra.run.vm09:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-10T05:53:54.341 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records in 2026-03-10T05:53:54.341 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records out 2026-03-10T05:53:54.341 INFO:teuthology.orchestra.run.vm09.stderr:512 bytes copied, 0.000178103 s, 2.9 MB/s 2026-03-10T05:53:54.342 DEBUG:teuthology.orchestra.run.vm09:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-10T05:53:54.402 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph orch apply osd --all-available-devices 2026-03-10T05:53:54.459 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:54 vm09 ceph-mon[53707]: Reconfiguring ceph-exporter.vm01 (monmap changed)... 2026-03-10T05:53:54.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:54 vm09 ceph-mon[53707]: Reconfiguring daemon ceph-exporter.vm01 on vm01 2026-03-10T05:53:54.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:54 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:54.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:54 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:54.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:54 vm09 ceph-mon[53707]: Reconfiguring crash.vm01 (monmap changed)... 2026-03-10T05:53:54.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:54 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm01", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-10T05:53:54.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:54 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:54.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:54 vm09 ceph-mon[53707]: Reconfiguring daemon crash.vm01 on vm01 2026-03-10T05:53:54.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:54 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:54.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:54 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:54.465 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:54 vm01 ceph-mon[47299]: Reconfiguring ceph-exporter.vm01 (monmap changed)... 2026-03-10T05:53:54.465 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:54 vm01 ceph-mon[47299]: Reconfiguring daemon ceph-exporter.vm01 on vm01 2026-03-10T05:53:54.465 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:54 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:54.465 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:54 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:54.465 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:54 vm01 ceph-mon[47299]: Reconfiguring crash.vm01 (monmap changed)... 2026-03-10T05:53:54.465 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:54 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm01", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-10T05:53:54.465 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:54 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:54.465 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:54 vm01 ceph-mon[47299]: Reconfiguring daemon crash.vm01 on vm01 2026-03-10T05:53:54.465 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:54 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:54.465 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:54 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:54.612 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm09/config 2026-03-10T05:53:54.887 INFO:teuthology.orchestra.run.vm09.stdout:Scheduled osd.all-available-devices update... 2026-03-10T05:53:55.033 INFO:tasks.cephadm:Waiting for 8 OSDs to come up... 2026-03-10T05:53:55.033 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd stat -f json 2026-03-10T05:53:55.354 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:53:55.593 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:55 vm01 ceph-mon[47299]: Reconfiguring alertmanager.vm01 (dependencies changed)... 2026-03-10T05:53:55.593 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:55 vm01 ceph-mon[47299]: Reconfiguring daemon alertmanager.vm01 on vm01 2026-03-10T05:53:55.593 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:55 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:55.593 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:55 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:55.593 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:55 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:55.610 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:53:55.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:55 vm09 ceph-mon[53707]: Reconfiguring alertmanager.vm01 (dependencies changed)... 2026-03-10T05:53:55.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:55 vm09 ceph-mon[53707]: Reconfiguring daemon alertmanager.vm01 on vm01 2026-03-10T05:53:55.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:55 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:55.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:55 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:55.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:55 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:55.783 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":5,"num_osds":0,"num_up_osds":0,"osd_up_since":0,"num_in_osds":0,"osd_in_since":0,"num_remapped_pgs":0} 2026-03-10T05:53:56.440 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:56 vm01 ceph-mon[47299]: Reconfiguring grafana.vm01 (dependencies changed)... 2026-03-10T05:53:56.440 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:56 vm01 ceph-mon[47299]: from='client.14254 -' entity='client.admin' cmd=[{"prefix": "orch apply osd", "all_available_devices": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:56.440 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:56 vm01 ceph-mon[47299]: Marking host: vm01 for OSDSpec preview refresh. 2026-03-10T05:53:56.440 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:56 vm01 ceph-mon[47299]: Marking host: vm09 for OSDSpec preview refresh. 2026-03-10T05:53:56.440 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:56 vm01 ceph-mon[47299]: Saving service osd.all-available-devices spec with placement * 2026-03-10T05:53:56.440 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:56 vm01 ceph-mon[47299]: Reconfiguring daemon grafana.vm01 on vm01 2026-03-10T05:53:56.440 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:56 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/1472715946' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:53:56.440 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:56 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:56.440 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:56 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:56.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:56 vm09 ceph-mon[53707]: Reconfiguring grafana.vm01 (dependencies changed)... 2026-03-10T05:53:56.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:56 vm09 ceph-mon[53707]: from='client.14254 -' entity='client.admin' cmd=[{"prefix": "orch apply osd", "all_available_devices": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:53:56.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:56 vm09 ceph-mon[53707]: Marking host: vm01 for OSDSpec preview refresh. 2026-03-10T05:53:56.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:56 vm09 ceph-mon[53707]: Marking host: vm09 for OSDSpec preview refresh. 2026-03-10T05:53:56.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:56 vm09 ceph-mon[53707]: Saving service osd.all-available-devices spec with placement * 2026-03-10T05:53:56.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:56 vm09 ceph-mon[53707]: Reconfiguring daemon grafana.vm01 on vm01 2026-03-10T05:53:56.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:56 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/1472715946' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:53:56.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:56 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:56.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:56 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:56.783 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd stat -f json 2026-03-10T05:53:57.095 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:53:57.358 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:53:57.513 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:57 vm01 ceph-mon[47299]: pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:53:57.513 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:57 vm01 ceph-mon[47299]: Reconfiguring prometheus.vm01 (dependencies changed)... 2026-03-10T05:53:57.513 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:57 vm01 ceph-mon[47299]: Reconfiguring daemon prometheus.vm01 on vm01 2026-03-10T05:53:57.513 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:57 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/2914443011' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:53:57.538 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":5,"num_osds":0,"num_up_osds":0,"osd_up_since":0,"num_in_osds":0,"osd_in_since":0,"num_remapped_pgs":0} 2026-03-10T05:53:57.635 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:57 vm09 ceph-mon[53707]: pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:53:57.635 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:57 vm09 ceph-mon[53707]: Reconfiguring prometheus.vm01 (dependencies changed)... 2026-03-10T05:53:57.635 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:57 vm09 ceph-mon[53707]: Reconfiguring daemon prometheus.vm01 on vm01 2026-03-10T05:53:57.635 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:57 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/2914443011' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:53:58.539 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd stat -f json 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:58 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:58 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:58 vm09 ceph-mon[53707]: Reconfiguring ceph-exporter.vm09 (monmap changed)... 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:58 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm09", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:58 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:58 vm09 ceph-mon[53707]: Reconfiguring daemon ceph-exporter.vm09 on vm09 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:58 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:58 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:58 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm09", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:58 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:58 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:58 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:58 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm09.uxbryv", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:58 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:58 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:58 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:58 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:58 vm01 ceph-mon[47299]: Reconfiguring ceph-exporter.vm09 (monmap changed)... 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:58 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm09", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:58 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:58 vm01 ceph-mon[47299]: Reconfiguring daemon ceph-exporter.vm09 on vm09 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:58 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:58 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:58 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm09", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:58 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:58 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:58 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:58 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm09.uxbryv", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T05:53:58.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:58 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T05:53:58.711 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:58 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:58.717 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:53:58.948 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:53:59.106 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":5,"num_osds":0,"num_up_osds":0,"osd_up_since":0,"num_in_osds":0,"osd_in_since":0,"num_remapped_pgs":0} 2026-03-10T05:53:59.563 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:59 vm09 ceph-mon[53707]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:53:59.563 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:59 vm09 ceph-mon[53707]: Reconfiguring crash.vm09 (monmap changed)... 2026-03-10T05:53:59.563 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:59 vm09 ceph-mon[53707]: Reconfiguring daemon crash.vm09 on vm09 2026-03-10T05:53:59.628 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:59 vm01 ceph-mon[47299]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:53:59.628 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:59 vm01 ceph-mon[47299]: Reconfiguring crash.vm09 (monmap changed)... 2026-03-10T05:53:59.628 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:59 vm01 ceph-mon[47299]: Reconfiguring daemon crash.vm09 on vm09 2026-03-10T05:53:59.628 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:59 vm01 ceph-mon[47299]: Reconfiguring mgr.vm09.uxbryv (monmap changed)... 2026-03-10T05:53:59.628 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:59 vm01 ceph-mon[47299]: Reconfiguring daemon mgr.vm09.uxbryv on vm09 2026-03-10T05:53:59.628 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:59 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:59.628 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:59 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:59.628 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:59 vm01 ceph-mon[47299]: Reconfiguring mon.vm09 (monmap changed)... 2026-03-10T05:53:59.628 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:59 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T05:53:59.628 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:59 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T05:53:59.628 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:59 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:59.628 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:59 vm01 ceph-mon[47299]: Reconfiguring daemon mon.vm09 on vm09 2026-03-10T05:53:59.628 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:59 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:59.628 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:59 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:59.628 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:59 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/1546384170' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:53:59.628 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:59 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-10T05:53:59.628 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:59 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm01.local:9093"}]: dispatch 2026-03-10T05:53:59.628 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:59 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:59.628 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:59 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-10T05:53:59.628 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:59 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm01.local:3000"}]: dispatch 2026-03-10T05:53:59.628 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:59 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:59.628 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:59 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-10T05:53:59.629 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:59 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm01.local:9095"}]: dispatch 2026-03-10T05:53:59.629 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:59 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:59.629 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:53:59 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:53:59.956 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:59 vm09 ceph-mon[53707]: Reconfiguring mgr.vm09.uxbryv (monmap changed)... 2026-03-10T05:53:59.956 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:59 vm09 ceph-mon[53707]: Reconfiguring daemon mgr.vm09.uxbryv on vm09 2026-03-10T05:53:59.956 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:59 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:59.956 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:59 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:59.956 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:59 vm09 ceph-mon[53707]: Reconfiguring mon.vm09 (monmap changed)... 2026-03-10T05:53:59.956 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:59 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T05:53:59.956 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:59 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T05:53:59.956 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:59 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:53:59.956 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:59 vm09 ceph-mon[53707]: Reconfiguring daemon mon.vm09 on vm09 2026-03-10T05:53:59.956 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:59 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:59.956 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:59 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:59.956 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:59 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/1546384170' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:53:59.956 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:59 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-10T05:53:59.956 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:59 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm01.local:9093"}]: dispatch 2026-03-10T05:53:59.956 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:59 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:59.956 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:59 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-10T05:53:59.956 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:59 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm01.local:3000"}]: dispatch 2026-03-10T05:53:59.956 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:59 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:59.956 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:59 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-10T05:53:59.956 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:59 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm01.local:9095"}]: dispatch 2026-03-10T05:53:59.956 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:59 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:53:59.956 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:53:59 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:54:00.107 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd stat -f json 2026-03-10T05:54:00.308 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:00.584 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:54:00.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:00 vm01 ceph-mon[47299]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-10T05:54:00.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:00 vm01 ceph-mon[47299]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm01.local:9093"}]: dispatch 2026-03-10T05:54:00.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:00 vm01 ceph-mon[47299]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-10T05:54:00.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:00 vm01 ceph-mon[47299]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm01.local:3000"}]: dispatch 2026-03-10T05:54:00.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:00 vm01 ceph-mon[47299]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-10T05:54:00.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:00 vm01 ceph-mon[47299]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm01.local:9095"}]: dispatch 2026-03-10T05:54:00.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:00 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:00.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:00 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:00.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:00 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:00.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:00 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:00.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:00 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:00.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:00 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:00.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:00 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:00.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:00 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:00.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:00 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:00.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:00 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:54:00.711 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:00 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:00.711 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:00 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:54:00.711 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:00 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T05:54:00.711 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:00 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:00.711 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:00 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T05:54:00.711 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:00 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:00.711 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:00 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/1185612922' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:00.749 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":5,"num_osds":0,"num_up_osds":0,"osd_up_since":0,"num_in_osds":0,"osd_in_since":0,"num_remapped_pgs":0} 2026-03-10T05:54:00.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:00 vm09 ceph-mon[53707]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-10T05:54:00.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:00 vm09 ceph-mon[53707]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm01.local:9093"}]: dispatch 2026-03-10T05:54:00.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:00 vm09 ceph-mon[53707]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-10T05:54:00.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:00 vm09 ceph-mon[53707]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm01.local:3000"}]: dispatch 2026-03-10T05:54:00.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:00 vm09 ceph-mon[53707]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-10T05:54:00.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:00 vm09 ceph-mon[53707]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm01.local:9095"}]: dispatch 2026-03-10T05:54:00.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:00 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:00.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:00 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:00.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:00 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:00.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:00 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:00.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:00 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:00.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:00 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:00.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:00 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:00.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:00 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:00.961 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:00 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:00.961 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:00 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:54:00.961 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:00 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:00.961 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:00 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:54:00.961 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:00 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T05:54:00.961 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:00 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:00.961 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:00 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T05:54:00.961 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:00 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:00.961 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:00 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/1185612922' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:01.751 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd stat -f json 2026-03-10T05:54:01.959 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:01 vm09 ceph-mon[53707]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:54:01.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:01 vm01 ceph-mon[47299]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:54:01.970 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:02.219 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:54:02.379 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":7,"num_osds":2,"num_up_osds":0,"osd_up_since":0,"num_in_osds":2,"osd_in_since":1773122041,"num_remapped_pgs":0} 2026-03-10T05:54:02.668 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:02 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/430056307' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a48f3e69-2204-46a2-aa6e-22512319163b"}]: dispatch 2026-03-10T05:54:02.959 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:02 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/430056307' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a48f3e69-2204-46a2-aa6e-22512319163b"}]: dispatch 2026-03-10T05:54:02.959 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:02 vm09 ceph-mon[53707]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a48f3e69-2204-46a2-aa6e-22512319163b"}]: dispatch 2026-03-10T05:54:02.959 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:02 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/2941439916' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "85e73c81-a2da-499a-8cf1-5da75fea2dd5"}]: dispatch 2026-03-10T05:54:02.959 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:02 vm09 ceph-mon[53707]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "a48f3e69-2204-46a2-aa6e-22512319163b"}]': finished 2026-03-10T05:54:02.959 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:02 vm09 ceph-mon[53707]: osdmap e6: 1 total, 0 up, 1 in 2026-03-10T05:54:02.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:02 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:54:02.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:02 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/2941439916' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "85e73c81-a2da-499a-8cf1-5da75fea2dd5"}]': finished 2026-03-10T05:54:02.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:02 vm09 ceph-mon[53707]: osdmap e7: 2 total, 0 up, 2 in 2026-03-10T05:54:02.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:02 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:54:02.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:02 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:54:02.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:02 vm09 ceph-mon[53707]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:54:02.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:02 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/187860024' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:54:02.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:02 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/773815619' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:54:02.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:02 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/4071689696' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:02.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:02 vm01 ceph-mon[47299]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a48f3e69-2204-46a2-aa6e-22512319163b"}]: dispatch 2026-03-10T05:54:02.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:02 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/2941439916' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "85e73c81-a2da-499a-8cf1-5da75fea2dd5"}]: dispatch 2026-03-10T05:54:02.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:02 vm01 ceph-mon[47299]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "a48f3e69-2204-46a2-aa6e-22512319163b"}]': finished 2026-03-10T05:54:02.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:02 vm01 ceph-mon[47299]: osdmap e6: 1 total, 0 up, 1 in 2026-03-10T05:54:02.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:02 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:54:02.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:02 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/2941439916' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "85e73c81-a2da-499a-8cf1-5da75fea2dd5"}]': finished 2026-03-10T05:54:02.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:02 vm01 ceph-mon[47299]: osdmap e7: 2 total, 0 up, 2 in 2026-03-10T05:54:02.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:02 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:54:02.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:02 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:54:02.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:02 vm01 ceph-mon[47299]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:54:02.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:02 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/187860024' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:54:02.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:02 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/773815619' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:54:02.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:02 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/4071689696' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:03.379 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd stat -f json 2026-03-10T05:54:03.537 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:03.777 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:54:03.960 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":7,"num_osds":2,"num_up_osds":0,"osd_up_since":0,"num_in_osds":2,"osd_in_since":1773122041,"num_remapped_pgs":0} 2026-03-10T05:54:04.828 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:04 vm09 ceph-mon[53707]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:54:04.828 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:04 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/188084210' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:04.950 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:04 vm01 ceph-mon[47299]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:54:04.950 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:04 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/188084210' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:04.960 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd stat -f json 2026-03-10T05:54:05.177 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:05.443 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:54:05.632 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":9,"num_osds":4,"num_up_osds":0,"osd_up_since":0,"num_in_osds":4,"osd_in_since":1773122045,"num_remapped_pgs":0} 2026-03-10T05:54:06.209 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:05 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/3708306996' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "441039ef-c8de-4d0a-aa8d-1950348d8d54"}]: dispatch 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:05 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/3708306996' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "441039ef-c8de-4d0a-aa8d-1950348d8d54"}]': finished 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:05 vm09 ceph-mon[53707]: osdmap e8: 3 total, 0 up, 3 in 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:05 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:05 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:05 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:05 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/1236615030' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "2cf13596-358c-4344-9953-0cc5ca8e025f"}]: dispatch 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:05 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/1236615030' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "2cf13596-358c-4344-9953-0cc5ca8e025f"}]': finished 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:05 vm09 ceph-mon[53707]: osdmap e9: 4 total, 0 up, 4 in 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:05 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:05 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:05 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:05 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:05 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/3460791847' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:05 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/3082876439' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:05 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/2742095017' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:05 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:05 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/3708306996' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "441039ef-c8de-4d0a-aa8d-1950348d8d54"}]: dispatch 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:05 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/3708306996' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "441039ef-c8de-4d0a-aa8d-1950348d8d54"}]': finished 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:05 vm01 ceph-mon[47299]: osdmap e8: 3 total, 0 up, 3 in 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:05 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:05 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:05 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:05 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/1236615030' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "2cf13596-358c-4344-9953-0cc5ca8e025f"}]: dispatch 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:05 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/1236615030' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "2cf13596-358c-4344-9953-0cc5ca8e025f"}]': finished 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:05 vm01 ceph-mon[47299]: osdmap e9: 4 total, 0 up, 4 in 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:05 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:05 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:05 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:05 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:05 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/3460791847' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:05 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/3082876439' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:05 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/2742095017' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:54:06.210 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:05 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T05:54:06.634 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd stat -f json 2026-03-10T05:54:06.831 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:07.063 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:54:07.174 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:07 vm01 ceph-mon[47299]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:54:07.174 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:07 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/1275045507' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:07.234 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":9,"num_osds":4,"num_up_osds":0,"osd_up_since":0,"num_in_osds":4,"osd_in_since":1773122045,"num_remapped_pgs":0} 2026-03-10T05:54:07.459 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:07 vm09 ceph-mon[53707]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:54:07.459 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:07 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/1275045507' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:08.235 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd stat -f json 2026-03-10T05:54:08.441 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:08.767 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:54:08.966 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":11,"num_osds":6,"num_up_osds":0,"osd_up_since":0,"num_in_osds":6,"osd_in_since":1773122048,"num_remapped_pgs":0} 2026-03-10T05:54:09.194 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:09 vm09 ceph-mon[53707]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:54:09.194 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:09 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/2572100164' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "b524239b-feb0-401f-bbd9-f924a65ba45c"}]: dispatch 2026-03-10T05:54:09.194 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:09 vm09 ceph-mon[53707]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "b524239b-feb0-401f-bbd9-f924a65ba45c"}]: dispatch 2026-03-10T05:54:09.194 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:09 vm09 ceph-mon[53707]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "b524239b-feb0-401f-bbd9-f924a65ba45c"}]': finished 2026-03-10T05:54:09.194 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:09 vm09 ceph-mon[53707]: osdmap e10: 5 total, 0 up, 5 in 2026-03-10T05:54:09.194 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:09 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:54:09.194 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:09 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:54:09.194 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:09 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:09.194 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:09 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:09.194 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:09 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:54:09.194 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:09 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/151347067' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "11ed13ca-a6e4-48db-9586-874dcc754728"}]: dispatch 2026-03-10T05:54:09.194 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:09 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/151347067' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "11ed13ca-a6e4-48db-9586-874dcc754728"}]': finished 2026-03-10T05:54:09.194 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:09 vm09 ceph-mon[53707]: osdmap e11: 6 total, 0 up, 6 in 2026-03-10T05:54:09.194 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:09 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:54:09.194 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:09 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:54:09.195 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:09 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:09.195 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:09 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:09.195 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:09 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:54:09.195 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:09 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:54:09.195 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:09 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/1275145363' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:09.195 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:09 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/459592885' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:54:09.195 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:09 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/1600861731' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:54:09.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:09 vm01 ceph-mon[47299]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:54:09.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:09 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/2572100164' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "b524239b-feb0-401f-bbd9-f924a65ba45c"}]: dispatch 2026-03-10T05:54:09.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:09 vm01 ceph-mon[47299]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "b524239b-feb0-401f-bbd9-f924a65ba45c"}]: dispatch 2026-03-10T05:54:09.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:09 vm01 ceph-mon[47299]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "b524239b-feb0-401f-bbd9-f924a65ba45c"}]': finished 2026-03-10T05:54:09.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:09 vm01 ceph-mon[47299]: osdmap e10: 5 total, 0 up, 5 in 2026-03-10T05:54:09.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:09 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:54:09.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:09 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:54:09.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:09 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:09.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:09 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:09.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:09 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:54:09.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:09 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/151347067' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "11ed13ca-a6e4-48db-9586-874dcc754728"}]: dispatch 2026-03-10T05:54:09.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:09 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/151347067' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "11ed13ca-a6e4-48db-9586-874dcc754728"}]': finished 2026-03-10T05:54:09.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:09 vm01 ceph-mon[47299]: osdmap e11: 6 total, 0 up, 6 in 2026-03-10T05:54:09.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:09 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:54:09.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:09 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:54:09.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:09 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:09.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:09 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:09.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:09 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:54:09.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:09 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:54:09.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:09 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/1275145363' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:09.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:09 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/459592885' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:54:09.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:09 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/1600861731' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:54:09.967 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd stat -f json 2026-03-10T05:54:10.133 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:10.370 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:54:10.518 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":11,"num_osds":6,"num_up_osds":0,"osd_up_since":0,"num_in_osds":6,"osd_in_since":1773122048,"num_remapped_pgs":0} 2026-03-10T05:54:11.459 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:11 vm09 ceph-mon[53707]: pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:54:11.459 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:11 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/3504010456' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:11.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:11 vm01 ceph-mon[47299]: pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:54:11.460 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:11 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/3504010456' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:11.519 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd stat -f json 2026-03-10T05:54:11.690 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:11.919 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:54:12.099 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":11,"num_osds":6,"num_up_osds":0,"osd_up_since":0,"num_in_osds":6,"osd_in_since":1773122048,"num_remapped_pgs":0} 2026-03-10T05:54:12.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:12 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/1805042192' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:12.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:12 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/474513336' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "f5910813-2e44-4da2-9c8e-48450f4a8fab"}]: dispatch 2026-03-10T05:54:12.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:12 vm01 ceph-mon[47299]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "f5910813-2e44-4da2-9c8e-48450f4a8fab"}]: dispatch 2026-03-10T05:54:12.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:12 vm01 ceph-mon[47299]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "f5910813-2e44-4da2-9c8e-48450f4a8fab"}]': finished 2026-03-10T05:54:12.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:12 vm01 ceph-mon[47299]: osdmap e12: 7 total, 0 up, 7 in 2026-03-10T05:54:12.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:12 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:54:12.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:12 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:54:12.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:12 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:12.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:12 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:12.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:12 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:54:12.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:12 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:54:12.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:12 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:54:12.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:12 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/3831070939' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "f4cbd2d8-a341-4674-88e2-75527b726469"}]: dispatch 2026-03-10T05:54:12.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:12 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/3831070939' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "f4cbd2d8-a341-4674-88e2-75527b726469"}]': finished 2026-03-10T05:54:12.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:12 vm01 ceph-mon[47299]: osdmap e13: 8 total, 0 up, 8 in 2026-03-10T05:54:12.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:12 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:54:12.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:12 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:54:12.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:12 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:12.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:12 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:12.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:12 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:54:12.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:12 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:54:12.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:12 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:54:12.355 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:12 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:54:12.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:12 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/1805042192' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:12.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:12 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/474513336' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "f5910813-2e44-4da2-9c8e-48450f4a8fab"}]: dispatch 2026-03-10T05:54:12.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:12 vm09 ceph-mon[53707]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "f5910813-2e44-4da2-9c8e-48450f4a8fab"}]: dispatch 2026-03-10T05:54:12.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:12 vm09 ceph-mon[53707]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "f5910813-2e44-4da2-9c8e-48450f4a8fab"}]': finished 2026-03-10T05:54:12.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:12 vm09 ceph-mon[53707]: osdmap e12: 7 total, 0 up, 7 in 2026-03-10T05:54:12.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:12 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:54:12.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:12 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:54:12.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:12 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:12.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:12 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:12.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:12 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:54:12.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:12 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:54:12.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:12 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:54:12.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:12 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/3831070939' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "f4cbd2d8-a341-4674-88e2-75527b726469"}]: dispatch 2026-03-10T05:54:12.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:12 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/3831070939' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "f4cbd2d8-a341-4674-88e2-75527b726469"}]': finished 2026-03-10T05:54:12.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:12 vm09 ceph-mon[53707]: osdmap e13: 8 total, 0 up, 8 in 2026-03-10T05:54:12.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:12 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:54:12.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:12 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:54:12.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:12 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:12.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:12 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:12.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:12 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:54:12.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:12 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:54:12.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:12 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:54:12.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:12 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:54:13.099 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd stat -f json 2026-03-10T05:54:13.274 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:13.390 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:13 vm01 ceph-mon[47299]: pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:54:13.391 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:13 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/1121213648' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:54:13.391 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:13 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/3310581248' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:54:13.459 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:13 vm09 ceph-mon[53707]: pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:54:13.459 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:13 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/1121213648' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:54:13.459 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:13 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/3310581248' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:54:13.511 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:54:13.687 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":13,"num_osds":8,"num_up_osds":0,"osd_up_since":0,"num_in_osds":8,"osd_in_since":1773122052,"num_remapped_pgs":0} 2026-03-10T05:54:14.688 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd stat -f json 2026-03-10T05:54:14.709 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:14 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/3283735501' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:14.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:14 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/3283735501' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:14.849 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:15.087 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:54:15.215 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:15 vm01 ceph-mon[47299]: pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:54:15.215 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:15 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/3343783575' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:15.255 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":13,"num_osds":8,"num_up_osds":0,"osd_up_since":0,"num_in_osds":8,"osd_in_since":1773122052,"num_remapped_pgs":0} 2026-03-10T05:54:15.554 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:15 vm09 ceph-mon[53707]: pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:54:15.554 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:15 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/3343783575' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:16.256 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd stat -f json 2026-03-10T05:54:16.467 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:16.771 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:54:16.948 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":13,"num_osds":8,"num_up_osds":0,"osd_up_since":0,"num_in_osds":8,"osd_in_since":1773122052,"num_remapped_pgs":0} 2026-03-10T05:54:17.299 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:17 vm09 ceph-mon[53707]: pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:54:17.299 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:17 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T05:54:17.299 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:17 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:17.299 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:17 vm09 ceph-mon[53707]: Deploying daemon osd.1 on vm01 2026-03-10T05:54:17.299 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:17 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T05:54:17.299 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:17 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:17.299 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:17 vm09 ceph-mon[53707]: Deploying daemon osd.0 on vm09 2026-03-10T05:54:17.299 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:17 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/3960665372' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:17.348 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:17 vm01 ceph-mon[47299]: pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:54:17.348 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:17 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T05:54:17.348 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:17 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:17.348 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:17 vm01 ceph-mon[47299]: Deploying daemon osd.1 on vm01 2026-03-10T05:54:17.348 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:17 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T05:54:17.348 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:17 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:17.348 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:17 vm01 ceph-mon[47299]: Deploying daemon osd.0 on vm09 2026-03-10T05:54:17.348 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:17 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/3960665372' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:17.950 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd stat -f json 2026-03-10T05:54:18.224 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:18.462 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:54:18.618 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":13,"num_osds":8,"num_up_osds":0,"osd_up_since":0,"num_in_osds":8,"osd_in_since":1773122052,"num_remapped_pgs":0} 2026-03-10T05:54:19.348 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:19 vm01 ceph-mon[47299]: pgmap v22: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:54:19.348 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:19 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/1695398072' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:19.348 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:19 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:19.348 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:19 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:19.348 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:19 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-10T05:54:19.348 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:19 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:19.348 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:19 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:19.348 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:19 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:19.348 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:19 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T05:54:19.348 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:19 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:19.447 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:19 vm09 ceph-mon[53707]: pgmap v22: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:54:19.447 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:19 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/1695398072' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:19.447 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:19 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:19.447 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:19 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:19.447 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:19 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-10T05:54:19.447 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:19 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:19.447 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:19 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:19.447 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:19 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:19.447 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:19 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T05:54:19.447 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:19 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:19.619 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd stat -f json 2026-03-10T05:54:19.869 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:20.300 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:20 vm09 ceph-mon[53707]: Deploying daemon osd.3 on vm01 2026-03-10T05:54:20.300 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:20 vm09 ceph-mon[53707]: Deploying daemon osd.2 on vm09 2026-03-10T05:54:20.312 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:54:20.475 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:20 vm01 ceph-mon[47299]: Deploying daemon osd.3 on vm01 2026-03-10T05:54:20.476 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:20 vm01 ceph-mon[47299]: Deploying daemon osd.2 on vm09 2026-03-10T05:54:20.517 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":13,"num_osds":8,"num_up_osds":0,"osd_up_since":0,"num_in_osds":8,"osd_in_since":1773122052,"num_remapped_pgs":0} 2026-03-10T05:54:21.402 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:21 vm09 ceph-mon[53707]: pgmap v23: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:54:21.402 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:21 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/2244660232' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:21.402 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:21 vm09 ceph-mon[53707]: from='osd.0 [v2:192.168.123.109:6800/3601312935,v1:192.168.123.109:6801/3601312935]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T05:54:21.402 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:21 vm09 ceph-mon[53707]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T05:54:21.402 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:21 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T05:54:21.402 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:21 vm09 ceph-mon[53707]: from='osd.1 [v2:192.168.123.101:6802/2835686876,v1:192.168.123.101:6803/2835686876]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T05:54:21.515 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:21 vm01 ceph-mon[47299]: pgmap v23: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:54:21.515 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:21 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/2244660232' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:21.515 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:21 vm01 ceph-mon[47299]: from='osd.0 [v2:192.168.123.109:6800/3601312935,v1:192.168.123.109:6801/3601312935]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T05:54:21.515 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:21 vm01 ceph-mon[47299]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T05:54:21.515 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:21 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T05:54:21.515 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:21 vm01 ceph-mon[47299]: from='osd.1 [v2:192.168.123.101:6802/2835686876,v1:192.168.123.101:6803/2835686876]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T05:54:21.518 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd stat -f json 2026-03-10T05:54:21.740 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:22.042 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:54:22.216 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":14,"num_osds":8,"num_up_osds":0,"osd_up_since":0,"num_in_osds":8,"osd_in_since":1773122052,"num_remapped_pgs":0} 2026-03-10T05:54:22.307 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:22 vm01 ceph-mon[47299]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-10T05:54:22.307 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:22 vm01 ceph-mon[47299]: from='osd.1 [v2:192.168.123.101:6802/2835686876,v1:192.168.123.101:6803/2835686876]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-10T05:54:22.307 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:22 vm01 ceph-mon[47299]: from='osd.0 [v2:192.168.123.109:6800/3601312935,v1:192.168.123.109:6801/3601312935]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T05:54:22.307 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:22 vm01 ceph-mon[47299]: osdmap e14: 8 total, 0 up, 8 in 2026-03-10T05:54:22.307 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:22 vm01 ceph-mon[47299]: from='osd.1 [v2:192.168.123.101:6802/2835686876,v1:192.168.123.101:6803/2835686876]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T05:54:22.307 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:22 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:54:22.307 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:22 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:54:22.307 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:22 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:22.307 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:22 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:22.307 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:22 vm01 ceph-mon[47299]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T05:54:22.307 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:22 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:54:22.307 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:22 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:54:22.307 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:22 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:54:22.307 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:22 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:54:22.307 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:22 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:22.307 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:22 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:22.307 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:22 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-10T05:54:22.307 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:22 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:22.307 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:22 vm01 ceph-mon[47299]: Deploying daemon osd.4 on vm09 2026-03-10T05:54:22.307 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:22 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:22.307 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:22 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:22.307 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:22 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-10T05:54:22.307 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:22 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:22.307 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:22 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/4104016301' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:22.548 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:22 vm09 ceph-mon[53707]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-10T05:54:22.548 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:22 vm09 ceph-mon[53707]: from='osd.1 [v2:192.168.123.101:6802/2835686876,v1:192.168.123.101:6803/2835686876]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-10T05:54:22.548 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:22 vm09 ceph-mon[53707]: from='osd.0 [v2:192.168.123.109:6800/3601312935,v1:192.168.123.109:6801/3601312935]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T05:54:22.548 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:22 vm09 ceph-mon[53707]: osdmap e14: 8 total, 0 up, 8 in 2026-03-10T05:54:22.548 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:22 vm09 ceph-mon[53707]: from='osd.1 [v2:192.168.123.101:6802/2835686876,v1:192.168.123.101:6803/2835686876]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T05:54:22.548 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:22 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:54:22.548 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:22 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:54:22.548 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:22 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:22.548 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:22 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:22.548 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:22 vm09 ceph-mon[53707]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T05:54:22.548 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:22 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:54:22.548 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:22 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:54:22.548 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:22 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:54:22.548 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:22 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:54:22.549 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:22 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:22.549 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:22 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:22.549 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:22 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-10T05:54:22.549 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:22 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:22.549 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:22 vm09 ceph-mon[53707]: Deploying daemon osd.4 on vm09 2026-03-10T05:54:22.549 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:22 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:22.549 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:22 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:22.549 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:22 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-10T05:54:22.549 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:22 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:22.549 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:22 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/4104016301' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:23.218 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd stat -f json 2026-03-10T05:54:23.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:23 vm09 ceph-mon[53707]: Deploying daemon osd.5 on vm01 2026-03-10T05:54:23.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:23 vm09 ceph-mon[53707]: pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:54:23.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:23 vm09 ceph-mon[53707]: from='osd.1 [v2:192.168.123.101:6802/2835686876,v1:192.168.123.101:6803/2835686876]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-10T05:54:23.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:23 vm09 ceph-mon[53707]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T05:54:23.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:23 vm09 ceph-mon[53707]: osdmap e15: 8 total, 0 up, 8 in 2026-03-10T05:54:23.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:23 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:54:23.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:23 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:54:23.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:23 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:23.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:23 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:23.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:23 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:54:23.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:23 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:54:23.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:23 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:54:23.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:23 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:54:23.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:23 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:54:23.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:23 vm09 ceph-mon[53707]: from='osd.3 [v2:192.168.123.101:6810/2800357583,v1:192.168.123.101:6811/2800357583]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-10T05:54:23.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:23 vm09 ceph-mon[53707]: from='osd.2 [v2:192.168.123.109:6808/2232889601,v1:192.168.123.109:6809/2232889601]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T05:54:23.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:23 vm09 ceph-mon[53707]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T05:54:23.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:23 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:54:23.460 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:23 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:54:23.484 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:23.537 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:23 vm01 ceph-mon[47299]: Deploying daemon osd.5 on vm01 2026-03-10T05:54:23.537 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:23 vm01 ceph-mon[47299]: pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:54:23.537 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:23 vm01 ceph-mon[47299]: from='osd.1 [v2:192.168.123.101:6802/2835686876,v1:192.168.123.101:6803/2835686876]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-10T05:54:23.537 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:23 vm01 ceph-mon[47299]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T05:54:23.537 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:23 vm01 ceph-mon[47299]: osdmap e15: 8 total, 0 up, 8 in 2026-03-10T05:54:23.537 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:23 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:54:23.538 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:23 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:54:23.538 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:23 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:23.538 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:23 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:23.538 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:23 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:54:23.538 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:23 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:54:23.538 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:23 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:54:23.538 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:23 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:54:23.538 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:23 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:54:23.538 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:23 vm01 ceph-mon[47299]: from='osd.3 [v2:192.168.123.101:6810/2800357583,v1:192.168.123.101:6811/2800357583]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-10T05:54:23.538 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:23 vm01 ceph-mon[47299]: from='osd.2 [v2:192.168.123.109:6808/2232889601,v1:192.168.123.109:6809/2232889601]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T05:54:23.538 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:23 vm01 ceph-mon[47299]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T05:54:23.538 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:23 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:54:23.538 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:23 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:54:23.875 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:54:24.060 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":16,"num_osds":8,"num_up_osds":0,"osd_up_since":0,"num_in_osds":8,"osd_in_since":1773122052,"num_remapped_pgs":0} 2026-03-10T05:54:24.363 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:24 vm09 ceph-mon[53707]: purged_snaps scrub starts 2026-03-10T05:54:24.363 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:24 vm09 ceph-mon[53707]: purged_snaps scrub ok 2026-03-10T05:54:24.363 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:24 vm09 ceph-mon[53707]: purged_snaps scrub starts 2026-03-10T05:54:24.363 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:24 vm09 ceph-mon[53707]: purged_snaps scrub ok 2026-03-10T05:54:24.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:24 vm09 ceph-mon[53707]: from='osd.3 [v2:192.168.123.101:6810/2800357583,v1:192.168.123.101:6811/2800357583]' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-10T05:54:24.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:24 vm09 ceph-mon[53707]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-10T05:54:24.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:24 vm09 ceph-mon[53707]: osdmap e16: 8 total, 0 up, 8 in 2026-03-10T05:54:24.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:24 vm09 ceph-mon[53707]: from='osd.2 [v2:192.168.123.109:6808/2232889601,v1:192.168.123.109:6809/2232889601]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T05:54:24.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:24 vm09 ceph-mon[53707]: from='osd.3 [v2:192.168.123.101:6810/2800357583,v1:192.168.123.101:6811/2800357583]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T05:54:24.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:24 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:54:24.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:24 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:54:24.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:24 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:24.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:24 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:24.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:24 vm09 ceph-mon[53707]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T05:54:24.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:24 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:54:24.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:24 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:54:24.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:24 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:54:24.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:24 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:54:24.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:24 vm09 ceph-mon[53707]: from='osd.0 ' entity='osd.0' 2026-03-10T05:54:24.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:24 vm09 ceph-mon[53707]: from='osd.1 [v2:192.168.123.101:6802/2835686876,v1:192.168.123.101:6803/2835686876]' entity='osd.1' 2026-03-10T05:54:24.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:24 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/3327026187' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:24.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:24 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:24.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:24 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:24.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:24 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-10T05:54:24.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:24 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:24.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:24 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:54:24.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:24 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:54:24.583 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:24 vm01 ceph-mon[47299]: purged_snaps scrub starts 2026-03-10T05:54:24.583 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:24 vm01 ceph-mon[47299]: purged_snaps scrub ok 2026-03-10T05:54:24.583 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:24 vm01 ceph-mon[47299]: purged_snaps scrub starts 2026-03-10T05:54:24.583 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:24 vm01 ceph-mon[47299]: purged_snaps scrub ok 2026-03-10T05:54:24.584 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:24 vm01 ceph-mon[47299]: from='osd.3 [v2:192.168.123.101:6810/2800357583,v1:192.168.123.101:6811/2800357583]' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-10T05:54:24.584 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:24 vm01 ceph-mon[47299]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-10T05:54:24.584 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:24 vm01 ceph-mon[47299]: osdmap e16: 8 total, 0 up, 8 in 2026-03-10T05:54:24.584 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:24 vm01 ceph-mon[47299]: from='osd.2 [v2:192.168.123.109:6808/2232889601,v1:192.168.123.109:6809/2232889601]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T05:54:24.584 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:24 vm01 ceph-mon[47299]: from='osd.3 [v2:192.168.123.101:6810/2800357583,v1:192.168.123.101:6811/2800357583]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T05:54:24.584 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:24 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:54:24.584 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:24 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:54:24.584 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:24 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:24.584 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:24 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:24.584 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:24 vm01 ceph-mon[47299]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T05:54:24.584 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:24 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:54:24.584 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:24 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:54:24.584 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:24 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:54:24.584 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:24 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:54:24.584 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:24 vm01 ceph-mon[47299]: from='osd.0 ' entity='osd.0' 2026-03-10T05:54:24.584 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:24 vm01 ceph-mon[47299]: from='osd.1 [v2:192.168.123.101:6802/2835686876,v1:192.168.123.101:6803/2835686876]' entity='osd.1' 2026-03-10T05:54:24.584 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:24 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/3327026187' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:24.584 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:24 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:24.584 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:24 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:24.584 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:24 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-10T05:54:24.584 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:24 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:24.584 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:24 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:54:24.584 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:24 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:54:25.061 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd stat -f json 2026-03-10T05:54:25.289 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:25.447 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:25 vm01 ceph-mon[47299]: pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:54:25.447 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:25 vm01 ceph-mon[47299]: Deploying daemon osd.6 on vm09 2026-03-10T05:54:25.447 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:25 vm01 ceph-mon[47299]: from='osd.3 [v2:192.168.123.101:6810/2800357583,v1:192.168.123.101:6811/2800357583]' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-10T05:54:25.447 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:25 vm01 ceph-mon[47299]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T05:54:25.447 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:25 vm01 ceph-mon[47299]: osd.0 [v2:192.168.123.109:6800/3601312935,v1:192.168.123.109:6801/3601312935] boot 2026-03-10T05:54:25.447 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:25 vm01 ceph-mon[47299]: osd.1 [v2:192.168.123.101:6802/2835686876,v1:192.168.123.101:6803/2835686876] boot 2026-03-10T05:54:25.447 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:25 vm01 ceph-mon[47299]: osdmap e17: 8 total, 2 up, 8 in 2026-03-10T05:54:25.447 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:25 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:54:25.447 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:25 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:54:25.447 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:25 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:25.447 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:25 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:25.447 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:25 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:54:25.447 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:25 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:54:25.447 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:25 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:54:25.447 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:25 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:54:25.447 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:25 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:25.448 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:25 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:25.448 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:25 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:25.448 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:25 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:25.448 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:25 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-10T05:54:25.448 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:25 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:25.565 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:25 vm09 ceph-mon[53707]: pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:54:25.565 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:25 vm09 ceph-mon[53707]: Deploying daemon osd.6 on vm09 2026-03-10T05:54:25.565 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:25 vm09 ceph-mon[53707]: from='osd.3 [v2:192.168.123.101:6810/2800357583,v1:192.168.123.101:6811/2800357583]' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-10T05:54:25.565 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:25 vm09 ceph-mon[53707]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T05:54:25.565 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:25 vm09 ceph-mon[53707]: osd.0 [v2:192.168.123.109:6800/3601312935,v1:192.168.123.109:6801/3601312935] boot 2026-03-10T05:54:25.565 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:25 vm09 ceph-mon[53707]: osd.1 [v2:192.168.123.101:6802/2835686876,v1:192.168.123.101:6803/2835686876] boot 2026-03-10T05:54:25.565 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:25 vm09 ceph-mon[53707]: osdmap e17: 8 total, 2 up, 8 in 2026-03-10T05:54:25.565 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:25 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:54:25.565 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:25 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:54:25.565 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:25 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:25.565 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:25 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:25.565 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:25 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:54:25.565 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:25 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:54:25.565 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:25 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:54:25.565 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:25 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:54:25.565 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:25 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:25.565 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:25 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:25.565 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:25 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:25.565 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:25 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:25.565 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:25 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-10T05:54:25.565 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:25 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:25.641 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:54:25.930 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":17,"num_osds":8,"num_up_osds":2,"osd_up_since":1773122064,"num_in_osds":8,"osd_in_since":1773122052,"num_remapped_pgs":0} 2026-03-10T05:54:26.677 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:26 vm09 ceph-mon[53707]: purged_snaps scrub starts 2026-03-10T05:54:26.677 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:26 vm09 ceph-mon[53707]: purged_snaps scrub ok 2026-03-10T05:54:26.677 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:26 vm09 ceph-mon[53707]: purged_snaps scrub starts 2026-03-10T05:54:26.677 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:26 vm09 ceph-mon[53707]: purged_snaps scrub ok 2026-03-10T05:54:26.677 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:26 vm09 ceph-mon[53707]: Deploying daemon osd.7 on vm01 2026-03-10T05:54:26.677 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:26 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/369243324' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:26.677 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:26 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:26.677 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:26 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:26.677 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:26 vm09 ceph-mon[53707]: from='osd.4 [v2:192.168.123.109:6816/1566157596,v1:192.168.123.109:6817/1566157596]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-10T05:54:26.677 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:26 vm09 ceph-mon[53707]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-10T05:54:26.677 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:26 vm09 ceph-mon[53707]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-10T05:54:26.677 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:26 vm09 ceph-mon[53707]: osdmap e18: 8 total, 2 up, 8 in 2026-03-10T05:54:26.677 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:26 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:26.677 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:26 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:26.677 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:26 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:54:26.677 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:26 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:54:26.677 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:26 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:54:26.677 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:26 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:54:26.677 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:26 vm09 ceph-mon[53707]: from='osd.4 [v2:192.168.123.109:6816/1566157596,v1:192.168.123.109:6817/1566157596]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T05:54:26.677 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:26 vm09 ceph-mon[53707]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T05:54:26.704 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:26 vm01 ceph-mon[47299]: purged_snaps scrub starts 2026-03-10T05:54:26.704 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:26 vm01 ceph-mon[47299]: purged_snaps scrub ok 2026-03-10T05:54:26.704 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:26 vm01 ceph-mon[47299]: purged_snaps scrub starts 2026-03-10T05:54:26.704 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:26 vm01 ceph-mon[47299]: purged_snaps scrub ok 2026-03-10T05:54:26.704 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:26 vm01 ceph-mon[47299]: Deploying daemon osd.7 on vm01 2026-03-10T05:54:26.704 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:26 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/369243324' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:26.704 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:26 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:26.704 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:26 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:26.704 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:26 vm01 ceph-mon[47299]: from='osd.4 [v2:192.168.123.109:6816/1566157596,v1:192.168.123.109:6817/1566157596]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-10T05:54:26.704 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:26 vm01 ceph-mon[47299]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-10T05:54:26.704 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:26 vm01 ceph-mon[47299]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-10T05:54:26.704 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:26 vm01 ceph-mon[47299]: osdmap e18: 8 total, 2 up, 8 in 2026-03-10T05:54:26.704 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:26 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:26.704 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:26 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:26.705 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:26 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:54:26.705 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:26 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:54:26.705 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:26 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:54:26.705 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:26 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:54:26.705 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:26 vm01 ceph-mon[47299]: from='osd.4 [v2:192.168.123.109:6816/1566157596,v1:192.168.123.109:6817/1566157596]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T05:54:26.705 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:26 vm01 ceph-mon[47299]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T05:54:26.931 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd stat -f json 2026-03-10T05:54:27.262 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:27.516 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:54:27.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:27 vm01 ceph-mon[47299]: pgmap v30: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-10T05:54:27.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:27 vm01 ceph-mon[47299]: from='osd.5 [v2:192.168.123.101:6818/1562465694,v1:192.168.123.101:6819/1562465694]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-10T05:54:27.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:27 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:27.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:27 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:27.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:27 vm01 ceph-mon[47299]: from='osd.2 ' entity='osd.2' 2026-03-10T05:54:27.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:27 vm01 ceph-mon[47299]: from='osd.3 [v2:192.168.123.101:6810/2800357583,v1:192.168.123.101:6811/2800357583]' entity='osd.3' 2026-03-10T05:54:27.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:27 vm01 ceph-mon[47299]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T05:54:27.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:27 vm01 ceph-mon[47299]: from='osd.5 [v2:192.168.123.101:6818/1562465694,v1:192.168.123.101:6819/1562465694]' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-10T05:54:27.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:27 vm01 ceph-mon[47299]: osd.2 [v2:192.168.123.109:6808/2232889601,v1:192.168.123.109:6809/2232889601] boot 2026-03-10T05:54:27.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:27 vm01 ceph-mon[47299]: osdmap e19: 8 total, 3 up, 8 in 2026-03-10T05:54:27.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:27 vm01 ceph-mon[47299]: from='osd.5 [v2:192.168.123.101:6818/1562465694,v1:192.168.123.101:6819/1562465694]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T05:54:27.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:27 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:27.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:27 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:27.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:27 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:54:27.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:27 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:54:27.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:27 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:54:27.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:27 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:54:27.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:27 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:54:27.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:27 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:27.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:27 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:27.671 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":19,"num_osds":8,"num_up_osds":3,"osd_up_since":1773122066,"num_in_osds":8,"osd_in_since":1773122052,"num_remapped_pgs":0} 2026-03-10T05:54:27.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:27 vm09 ceph-mon[53707]: pgmap v30: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-10T05:54:27.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:27 vm09 ceph-mon[53707]: from='osd.5 [v2:192.168.123.101:6818/1562465694,v1:192.168.123.101:6819/1562465694]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-10T05:54:27.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:27 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:27.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:27 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:27.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:27 vm09 ceph-mon[53707]: from='osd.2 ' entity='osd.2' 2026-03-10T05:54:27.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:27 vm09 ceph-mon[53707]: from='osd.3 [v2:192.168.123.101:6810/2800357583,v1:192.168.123.101:6811/2800357583]' entity='osd.3' 2026-03-10T05:54:27.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:27 vm09 ceph-mon[53707]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T05:54:27.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:27 vm09 ceph-mon[53707]: from='osd.5 [v2:192.168.123.101:6818/1562465694,v1:192.168.123.101:6819/1562465694]' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-10T05:54:27.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:27 vm09 ceph-mon[53707]: osd.2 [v2:192.168.123.109:6808/2232889601,v1:192.168.123.109:6809/2232889601] boot 2026-03-10T05:54:27.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:27 vm09 ceph-mon[53707]: osdmap e19: 8 total, 3 up, 8 in 2026-03-10T05:54:27.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:27 vm09 ceph-mon[53707]: from='osd.5 [v2:192.168.123.101:6818/1562465694,v1:192.168.123.101:6819/1562465694]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T05:54:27.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:27 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:54:27.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:27 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:27.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:27 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:54:27.711 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:27 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:54:27.711 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:27 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:54:27.711 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:27 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:54:27.711 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:27 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:54:27.711 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:27 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:27.711 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:27 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:28.672 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd stat -f json 2026-03-10T05:54:28.709 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:28 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/922201911' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:28.709 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:28 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:28 vm09 ceph-mon[53707]: from='osd.5 [v2:192.168.123.101:6818/1562465694,v1:192.168.123.101:6819/1562465694]' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:28 vm09 ceph-mon[53707]: osd.3 [v2:192.168.123.101:6810/2800357583,v1:192.168.123.101:6811/2800357583] boot 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:28 vm09 ceph-mon[53707]: osd.4 [v2:192.168.123.109:6816/1566157596,v1:192.168.123.109:6817/1566157596] boot 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:28 vm09 ceph-mon[53707]: osdmap e20: 8 total, 5 up, 8 in 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:28 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:28 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:28 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:28 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:28 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:28 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:28 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:28 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:28 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:28 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:28 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:28 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/922201911' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:28 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:28 vm01 ceph-mon[47299]: from='osd.5 [v2:192.168.123.101:6818/1562465694,v1:192.168.123.101:6819/1562465694]' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:28 vm01 ceph-mon[47299]: osd.3 [v2:192.168.123.101:6810/2800357583,v1:192.168.123.101:6811/2800357583] boot 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:28 vm01 ceph-mon[47299]: osd.4 [v2:192.168.123.109:6816/1566157596,v1:192.168.123.109:6817/1566157596] boot 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:28 vm01 ceph-mon[47299]: osdmap e20: 8 total, 5 up, 8 in 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:28 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:28 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:28 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:28 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:28 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:28 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:28 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:28 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:28 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:28 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:28.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:28 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:28.928 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:29.213 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:54:29.377 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":21,"num_osds":8,"num_up_osds":6,"osd_up_since":1773122068,"num_in_osds":8,"osd_in_since":1773122052,"num_remapped_pgs":0} 2026-03-10T05:54:29.603 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:29 vm01 ceph-mon[47299]: purged_snaps scrub starts 2026-03-10T05:54:29.603 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:29 vm01 ceph-mon[47299]: purged_snaps scrub ok 2026-03-10T05:54:29.603 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:29 vm01 ceph-mon[47299]: purged_snaps scrub starts 2026-03-10T05:54:29.603 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:29 vm01 ceph-mon[47299]: purged_snaps scrub ok 2026-03-10T05:54:29.603 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:29 vm01 ceph-mon[47299]: pgmap v33: 0 pgs: ; 0 B data, 879 MiB used, 59 GiB / 60 GiB avail 2026-03-10T05:54:29.603 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:29 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-10T05:54:29.603 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:29 vm01 ceph-mon[47299]: osd.5 [v2:192.168.123.101:6818/1562465694,v1:192.168.123.101:6819/1562465694] boot 2026-03-10T05:54:29.603 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:29 vm01 ceph-mon[47299]: osdmap e21: 8 total, 6 up, 8 in 2026-03-10T05:54:29.603 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:29 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:54:29.603 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:29 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:54:29.603 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:29 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:54:29.603 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:29 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T05:54:29.603 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:29 vm01 ceph-mon[47299]: from='osd.6 [v2:192.168.123.109:6824/1216154444,v1:192.168.123.109:6825/1216154444]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-10T05:54:29.603 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:29 vm01 ceph-mon[47299]: from='osd.6 ' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-10T05:54:29.603 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:29 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/2353358519' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:29.603 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:29 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:29.603 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:29 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:29.603 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:29 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:54:29.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:29 vm09 ceph-mon[53707]: purged_snaps scrub starts 2026-03-10T05:54:29.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:29 vm09 ceph-mon[53707]: purged_snaps scrub ok 2026-03-10T05:54:29.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:29 vm09 ceph-mon[53707]: purged_snaps scrub starts 2026-03-10T05:54:29.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:29 vm09 ceph-mon[53707]: purged_snaps scrub ok 2026-03-10T05:54:29.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:29 vm09 ceph-mon[53707]: pgmap v33: 0 pgs: ; 0 B data, 879 MiB used, 59 GiB / 60 GiB avail 2026-03-10T05:54:29.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:29 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-10T05:54:29.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:29 vm09 ceph-mon[53707]: osd.5 [v2:192.168.123.101:6818/1562465694,v1:192.168.123.101:6819/1562465694] boot 2026-03-10T05:54:29.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:29 vm09 ceph-mon[53707]: osdmap e21: 8 total, 6 up, 8 in 2026-03-10T05:54:29.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:29 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:54:29.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:29 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:54:29.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:29 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:54:29.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:29 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T05:54:29.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:29 vm09 ceph-mon[53707]: from='osd.6 [v2:192.168.123.109:6824/1216154444,v1:192.168.123.109:6825/1216154444]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-10T05:54:29.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:29 vm09 ceph-mon[53707]: from='osd.6 ' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-10T05:54:29.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:29 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/2353358519' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:29.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:29 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:29.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:29 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:29.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:29 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:54:30.378 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd stat -f json 2026-03-10T05:54:30.619 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:30.900 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:54:30.914 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:30 vm01 ceph-mon[47299]: pgmap v36: 1 pgs: 1 unknown; 0 B data, 1.3 GiB used, 99 GiB / 100 GiB avail 2026-03-10T05:54:30.914 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:30 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-10T05:54:30.914 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:30 vm01 ceph-mon[47299]: from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-10T05:54:30.914 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:30 vm01 ceph-mon[47299]: osdmap e22: 8 total, 6 up, 8 in 2026-03-10T05:54:30.914 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:30 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:54:30.914 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:30 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:54:30.914 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:30 vm01 ceph-mon[47299]: from='osd.6 [v2:192.168.123.109:6824/1216154444,v1:192.168.123.109:6825/1216154444]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T05:54:30.914 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:30 vm01 ceph-mon[47299]: from='osd.6 ' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T05:54:30.914 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:30 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:30.914 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:30 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:30.914 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:30 vm01 ceph-mon[47299]: from='osd.7 [v2:192.168.123.101:6826/1172902724,v1:192.168.123.101:6827/1172902724]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-10T05:54:30.914 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:30 vm01 ceph-mon[47299]: Health check failed: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T05:54:30.914 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:30 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:30.914 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:30 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:30.928 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:30 vm09 ceph-mon[53707]: pgmap v36: 1 pgs: 1 unknown; 0 B data, 1.3 GiB used, 99 GiB / 100 GiB avail 2026-03-10T05:54:30.928 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:30 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-10T05:54:30.928 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:30 vm09 ceph-mon[53707]: from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-10T05:54:30.928 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:30 vm09 ceph-mon[53707]: osdmap e22: 8 total, 6 up, 8 in 2026-03-10T05:54:30.928 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:30 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:54:30.928 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:30 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:54:30.928 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:30 vm09 ceph-mon[53707]: from='osd.6 [v2:192.168.123.109:6824/1216154444,v1:192.168.123.109:6825/1216154444]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T05:54:30.928 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:30 vm09 ceph-mon[53707]: from='osd.6 ' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T05:54:30.928 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:30 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:30.928 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:30 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:30.928 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:30 vm09 ceph-mon[53707]: from='osd.7 [v2:192.168.123.101:6826/1172902724,v1:192.168.123.101:6827/1172902724]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-10T05:54:30.928 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:30 vm09 ceph-mon[53707]: Health check failed: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T05:54:30.928 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:30 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:30.928 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:30 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:31.071 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":23,"num_osds":8,"num_up_osds":6,"osd_up_since":1773122068,"num_in_osds":8,"osd_in_since":1773122052,"num_remapped_pgs":0} 2026-03-10T05:54:31.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:31 vm01 ceph-mon[47299]: from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T05:54:31.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:31 vm01 ceph-mon[47299]: from='osd.7 [v2:192.168.123.101:6826/1172902724,v1:192.168.123.101:6827/1172902724]' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-10T05:54:31.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:31 vm01 ceph-mon[47299]: osdmap e23: 8 total, 6 up, 8 in 2026-03-10T05:54:31.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:31 vm01 ceph-mon[47299]: from='osd.7 [v2:192.168.123.101:6826/1172902724,v1:192.168.123.101:6827/1172902724]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T05:54:31.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:31 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:54:31.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:31 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:54:31.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:31 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/184533736' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:31.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:31 vm01 ceph-mon[47299]: Detected new or changed devices on vm09 2026-03-10T05:54:31.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:31 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:31.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:31 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:31.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:31 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:54:31.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:31 vm01 ceph-mon[47299]: from='osd.7 [v2:192.168.123.101:6826/1172902724,v1:192.168.123.101:6827/1172902724]' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-10T05:54:31.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:31 vm01 ceph-mon[47299]: osd.6 [v2:192.168.123.109:6824/1216154444,v1:192.168.123.109:6825/1216154444] boot 2026-03-10T05:54:31.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:31 vm01 ceph-mon[47299]: osdmap e24: 8 total, 7 up, 8 in 2026-03-10T05:54:31.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:31 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:54:31.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:31 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:54:31.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:31 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:32.072 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd stat -f json 2026-03-10T05:54:32.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:31 vm09 ceph-mon[53707]: from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T05:54:32.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:31 vm09 ceph-mon[53707]: from='osd.7 [v2:192.168.123.101:6826/1172902724,v1:192.168.123.101:6827/1172902724]' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-10T05:54:32.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:31 vm09 ceph-mon[53707]: osdmap e23: 8 total, 6 up, 8 in 2026-03-10T05:54:32.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:31 vm09 ceph-mon[53707]: from='osd.7 [v2:192.168.123.101:6826/1172902724,v1:192.168.123.101:6827/1172902724]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T05:54:32.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:31 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:54:32.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:31 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:54:32.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:31 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/184533736' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:32.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:31 vm09 ceph-mon[53707]: Detected new or changed devices on vm09 2026-03-10T05:54:32.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:31 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:32.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:31 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:32.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:31 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:54:32.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:31 vm09 ceph-mon[53707]: from='osd.7 [v2:192.168.123.101:6826/1172902724,v1:192.168.123.101:6827/1172902724]' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-10T05:54:32.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:31 vm09 ceph-mon[53707]: osd.6 [v2:192.168.123.109:6824/1216154444,v1:192.168.123.109:6825/1216154444] boot 2026-03-10T05:54:32.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:31 vm09 ceph-mon[53707]: osdmap e24: 8 total, 7 up, 8 in 2026-03-10T05:54:32.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:31 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:54:32.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:31 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:54:32.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:31 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:32.260 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:32.480 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:54:32.646 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":24,"num_osds":8,"num_up_osds":7,"osd_up_since":1773122071,"num_in_osds":8,"osd_in_since":1773122052,"num_remapped_pgs":1} 2026-03-10T05:54:32.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:32 vm01 ceph-mon[47299]: purged_snaps scrub starts 2026-03-10T05:54:32.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:32 vm01 ceph-mon[47299]: purged_snaps scrub ok 2026-03-10T05:54:32.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:32 vm01 ceph-mon[47299]: pgmap v39: 1 pgs: 1 unknown; 0 B data, 958 MiB used, 119 GiB / 120 GiB avail 2026-03-10T05:54:32.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:32 vm01 ceph-mon[47299]: Detected new or changed devices on vm01 2026-03-10T05:54:32.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:32 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:32.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:32 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:54:32.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:32 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:32.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:32 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:54:32.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:32 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:32.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:32 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:54:32.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:32 vm01 ceph-mon[47299]: Health check cleared: POOL_APP_NOT_ENABLED (was: 1 pool(s) do not have an application enabled) 2026-03-10T05:54:32.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:32 vm01 ceph-mon[47299]: Cluster is now healthy 2026-03-10T05:54:32.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:32 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/753727167' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:32.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:32 vm01 ceph-mon[47299]: osd.7 [v2:192.168.123.101:6826/1172902724,v1:192.168.123.101:6827/1172902724] boot 2026-03-10T05:54:32.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:32 vm01 ceph-mon[47299]: osdmap e25: 8 total, 8 up, 8 in 2026-03-10T05:54:32.960 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:32 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:54:33.209 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:32 vm09 ceph-mon[53707]: purged_snaps scrub starts 2026-03-10T05:54:33.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:32 vm09 ceph-mon[53707]: purged_snaps scrub ok 2026-03-10T05:54:33.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:32 vm09 ceph-mon[53707]: pgmap v39: 1 pgs: 1 unknown; 0 B data, 958 MiB used, 119 GiB / 120 GiB avail 2026-03-10T05:54:33.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:32 vm09 ceph-mon[53707]: Detected new or changed devices on vm01 2026-03-10T05:54:33.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:32 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:33.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:32 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:54:33.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:32 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:33.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:32 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:54:33.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:32 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:33.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:32 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:54:33.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:32 vm09 ceph-mon[53707]: Health check cleared: POOL_APP_NOT_ENABLED (was: 1 pool(s) do not have an application enabled) 2026-03-10T05:54:33.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:32 vm09 ceph-mon[53707]: Cluster is now healthy 2026-03-10T05:54:33.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:32 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/753727167' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:33.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:32 vm09 ceph-mon[53707]: osd.7 [v2:192.168.123.101:6826/1172902724,v1:192.168.123.101:6827/1172902724] boot 2026-03-10T05:54:33.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:32 vm09 ceph-mon[53707]: osdmap e25: 8 total, 8 up, 8 in 2026-03-10T05:54:33.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:32 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:54:33.647 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd stat -f json 2026-03-10T05:54:33.834 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:33.950 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:33 vm01 ceph-mon[47299]: purged_snaps scrub starts 2026-03-10T05:54:33.950 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:33 vm01 ceph-mon[47299]: purged_snaps scrub ok 2026-03-10T05:54:33.950 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:33 vm01 ceph-mon[47299]: osdmap e26: 8 total, 8 up, 8 in 2026-03-10T05:54:34.063 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:54:34.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:33 vm09 ceph-mon[53707]: purged_snaps scrub starts 2026-03-10T05:54:34.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:33 vm09 ceph-mon[53707]: purged_snaps scrub ok 2026-03-10T05:54:34.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:33 vm09 ceph-mon[53707]: osdmap e26: 8 total, 8 up, 8 in 2026-03-10T05:54:34.211 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":26,"num_osds":8,"num_up_osds":8,"osd_up_since":1773122072,"num_in_osds":8,"osd_in_since":1773122052,"num_remapped_pgs":0} 2026-03-10T05:54:34.212 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd dump --format=json 2026-03-10T05:54:34.373 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:34.600 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:54:34.600 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":26,"fsid":"100dc3f2-1c45-11f1-8010-b96ca19c2df4","created":"2026-03-10T05:52:41.180004+0000","modified":"2026-03-10T05:54:33.732036+0000","last_up_change":"2026-03-10T05:54:32.729144+0000","last_in_change":"2026-03-10T05:54:12.101073+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":12,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-10T05:54:27.730315+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"23","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":7.8899998664855957,"score_stable":7.8899998664855957,"optimal_score":0.37999999523162842,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"a48f3e69-2204-46a2-aa6e-22512319163b","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6800","nonce":3601312935},{"type":"v1","addr":"192.168.123.109:6801","nonce":3601312935}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6802","nonce":3601312935},{"type":"v1","addr":"192.168.123.109:6803","nonce":3601312935}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6806","nonce":3601312935},{"type":"v1","addr":"192.168.123.109:6807","nonce":3601312935}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6804","nonce":3601312935},{"type":"v1","addr":"192.168.123.109:6805","nonce":3601312935}]},"public_addr":"192.168.123.109:6801/3601312935","cluster_addr":"192.168.123.109:6803/3601312935","heartbeat_back_addr":"192.168.123.109:6807/3601312935","heartbeat_front_addr":"192.168.123.109:6805/3601312935","state":["exists","up"]},{"osd":1,"uuid":"85e73c81-a2da-499a-8cf1-5da75fea2dd5","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6802","nonce":2835686876},{"type":"v1","addr":"192.168.123.101:6803","nonce":2835686876}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6804","nonce":2835686876},{"type":"v1","addr":"192.168.123.101:6805","nonce":2835686876}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6808","nonce":2835686876},{"type":"v1","addr":"192.168.123.101:6809","nonce":2835686876}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6806","nonce":2835686876},{"type":"v1","addr":"192.168.123.101:6807","nonce":2835686876}]},"public_addr":"192.168.123.101:6803/2835686876","cluster_addr":"192.168.123.101:6805/2835686876","heartbeat_back_addr":"192.168.123.101:6809/2835686876","heartbeat_front_addr":"192.168.123.101:6807/2835686876","state":["exists","up"]},{"osd":2,"uuid":"441039ef-c8de-4d0a-aa8d-1950348d8d54","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":19,"up_thru":21,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6808","nonce":2232889601},{"type":"v1","addr":"192.168.123.109:6809","nonce":2232889601}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6810","nonce":2232889601},{"type":"v1","addr":"192.168.123.109:6811","nonce":2232889601}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6814","nonce":2232889601},{"type":"v1","addr":"192.168.123.109:6815","nonce":2232889601}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6812","nonce":2232889601},{"type":"v1","addr":"192.168.123.109:6813","nonce":2232889601}]},"public_addr":"192.168.123.109:6809/2232889601","cluster_addr":"192.168.123.109:6811/2232889601","heartbeat_back_addr":"192.168.123.109:6815/2232889601","heartbeat_front_addr":"192.168.123.109:6813/2232889601","state":["exists","up"]},{"osd":3,"uuid":"2cf13596-358c-4344-9953-0cc5ca8e025f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":20,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6810","nonce":2800357583},{"type":"v1","addr":"192.168.123.101:6811","nonce":2800357583}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6812","nonce":2800357583},{"type":"v1","addr":"192.168.123.101:6813","nonce":2800357583}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6816","nonce":2800357583},{"type":"v1","addr":"192.168.123.101:6817","nonce":2800357583}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6814","nonce":2800357583},{"type":"v1","addr":"192.168.123.101:6815","nonce":2800357583}]},"public_addr":"192.168.123.101:6811/2800357583","cluster_addr":"192.168.123.101:6813/2800357583","heartbeat_back_addr":"192.168.123.101:6817/2800357583","heartbeat_front_addr":"192.168.123.101:6815/2800357583","state":["exists","up"]},{"osd":4,"uuid":"b524239b-feb0-401f-bbd9-f924a65ba45c","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":20,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6816","nonce":1566157596},{"type":"v1","addr":"192.168.123.109:6817","nonce":1566157596}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6818","nonce":1566157596},{"type":"v1","addr":"192.168.123.109:6819","nonce":1566157596}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6822","nonce":1566157596},{"type":"v1","addr":"192.168.123.109:6823","nonce":1566157596}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6820","nonce":1566157596},{"type":"v1","addr":"192.168.123.109:6821","nonce":1566157596}]},"public_addr":"192.168.123.109:6817/1566157596","cluster_addr":"192.168.123.109:6819/1566157596","heartbeat_back_addr":"192.168.123.109:6823/1566157596","heartbeat_front_addr":"192.168.123.109:6821/1566157596","state":["exists","up"]},{"osd":5,"uuid":"11ed13ca-a6e4-48db-9586-874dcc754728","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":21,"up_thru":24,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6818","nonce":1562465694},{"type":"v1","addr":"192.168.123.101:6819","nonce":1562465694}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6820","nonce":1562465694},{"type":"v1","addr":"192.168.123.101:6821","nonce":1562465694}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6824","nonce":1562465694},{"type":"v1","addr":"192.168.123.101:6825","nonce":1562465694}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6822","nonce":1562465694},{"type":"v1","addr":"192.168.123.101:6823","nonce":1562465694}]},"public_addr":"192.168.123.101:6819/1562465694","cluster_addr":"192.168.123.101:6821/1562465694","heartbeat_back_addr":"192.168.123.101:6825/1562465694","heartbeat_front_addr":"192.168.123.101:6823/1562465694","state":["exists","up"]},{"osd":6,"uuid":"f5910813-2e44-4da2-9c8e-48450f4a8fab","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":24,"up_thru":25,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6824","nonce":1216154444},{"type":"v1","addr":"192.168.123.109:6825","nonce":1216154444}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6826","nonce":1216154444},{"type":"v1","addr":"192.168.123.109:6827","nonce":1216154444}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6830","nonce":1216154444},{"type":"v1","addr":"192.168.123.109:6831","nonce":1216154444}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6828","nonce":1216154444},{"type":"v1","addr":"192.168.123.109:6829","nonce":1216154444}]},"public_addr":"192.168.123.109:6825/1216154444","cluster_addr":"192.168.123.109:6827/1216154444","heartbeat_back_addr":"192.168.123.109:6831/1216154444","heartbeat_front_addr":"192.168.123.109:6829/1216154444","state":["exists","up"]},{"osd":7,"uuid":"f4cbd2d8-a341-4674-88e2-75527b726469","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":25,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6826","nonce":1172902724},{"type":"v1","addr":"192.168.123.101:6827","nonce":1172902724}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6828","nonce":1172902724},{"type":"v1","addr":"192.168.123.101:6829","nonce":1172902724}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6832","nonce":1172902724},{"type":"v1","addr":"192.168.123.101:6833","nonce":1172902724}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6830","nonce":1172902724},{"type":"v1","addr":"192.168.123.101:6831","nonce":1172902724}]},"public_addr":"192.168.123.101:6827/1172902724","cluster_addr":"192.168.123.101:6829/1172902724","heartbeat_back_addr":"192.168.123.101:6833/1172902724","heartbeat_front_addr":"192.168.123.101:6831/1172902724","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:54:21.619546+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:54:21.765397+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:54:24.139079+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:54:24.043381+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:54:26.722231+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:54:27.434731+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:54:29.836722+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:54:31.300808+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.101:6801/1561074913":"2026-03-11T05:53:35.455537+0000","192.168.123.101:0/877175877":"2026-03-11T05:53:35.455537+0000","192.168.123.101:0/100373818":"2026-03-11T05:53:35.455537+0000","192.168.123.101:0/3534784264":"2026-03-11T05:53:05.831349+0000","192.168.123.101:6800/1229025911":"2026-03-11T05:52:53.419293+0000","192.168.123.101:6800/1371156494":"2026-03-11T05:53:05.831349+0000","192.168.123.101:0/500832301":"2026-03-11T05:52:53.419293+0000","192.168.123.101:6801/1229025911":"2026-03-11T05:52:53.419293+0000","192.168.123.101:6801/1371156494":"2026-03-11T05:53:05.831349+0000","192.168.123.101:0/3827013041":"2026-03-11T05:52:53.419293+0000","192.168.123.101:0/2988507782":"2026-03-11T05:52:53.419293+0000","192.168.123.101:0/1403556378":"2026-03-11T05:53:05.831349+0000","192.168.123.101:0/2020700462":"2026-03-11T05:53:05.831349+0000","192.168.123.101:6800/1561074913":"2026-03-11T05:53:35.455537+0000","192.168.123.101:0/1369389137":"2026-03-11T05:53:35.455537+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-10T05:54:34.770 INFO:tasks.cephadm.ceph_manager.ceph:[{'pool': 1, 'pool_name': '.mgr', 'create_time': '2026-03-10T05:54:27.730315+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'is_stretch_pool': False, 'object_hash': 2, 'pg_autoscale_mode': 'off', 'pg_num': 1, 'pg_placement_num': 1, 'pg_placement_num_target': 1, 'pg_num_target': 1, 'pg_num_pending': 1, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': "0'0", 'target_version': "0'0"}, 'last_change': '23', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {'pg_num_max': 32, 'pg_num_min': 1}, 'application_metadata': {'mgr': {}}, 'read_balance': {'score_type': 'Fair distribution', 'score_acting': 7.889999866485596, 'score_stable': 7.889999866485596, 'optimal_score': 0.3799999952316284, 'raw_score_acting': 3, 'raw_score_stable': 3, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}}] 2026-03-10T05:54:34.770 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd pool get .mgr pg_num 2026-03-10T05:54:34.930 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:34.997 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:34 vm01 ceph-mon[47299]: pgmap v42: 1 pgs: 1 unknown; 0 B data, 211 MiB used, 160 GiB / 160 GiB avail 2026-03-10T05:54:34.997 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:34 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/3709604844' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:34.997 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:34 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/3973142957' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:54:35.137 INFO:teuthology.orchestra.run.vm01.stdout:pg_num: 1 2026-03-10T05:54:35.209 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:34 vm09 ceph-mon[53707]: pgmap v42: 1 pgs: 1 unknown; 0 B data, 211 MiB used, 160 GiB / 160 GiB avail 2026-03-10T05:54:35.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:34 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/3709604844' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:54:35.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:34 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/3973142957' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:54:35.288 INFO:tasks.cephadm:Setting up client nodes... 2026-03-10T05:54:35.288 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph auth get-or-create client.0 mon 'allow *' osd 'allow *' mds 'allow *' mgr 'allow *' 2026-03-10T05:54:35.446 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:35.709 INFO:teuthology.orchestra.run.vm01.stdout:[client.0] 2026-03-10T05:54:35.709 INFO:teuthology.orchestra.run.vm01.stdout: key = AQAbsq9pSjIRKhAAOsbxgHTt8U5VhDSICEz9pQ== 2026-03-10T05:54:35.848 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:35 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/149904811' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-10T05:54:35.848 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:35 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/3113882562' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T05:54:35.848 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:35 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/3113882562' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-10T05:54:35.848 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:35 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T05:54:35.873 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T05:54:35.873 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/ceph/ceph.client.0.keyring 2026-03-10T05:54:35.873 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod 0644 /etc/ceph/ceph.client.0.keyring 2026-03-10T05:54:35.905 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph auth get-or-create client.1 mon 'allow *' osd 'allow *' mds 'allow *' mgr 'allow *' 2026-03-10T05:54:36.073 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm09/config 2026-03-10T05:54:36.097 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:35 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/149904811' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-10T05:54:36.097 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:35 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/3113882562' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T05:54:36.097 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:35 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/3113882562' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-10T05:54:36.097 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:35 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T05:54:36.344 INFO:teuthology.orchestra.run.vm09.stdout:[client.1] 2026-03-10T05:54:36.344 INFO:teuthology.orchestra.run.vm09.stdout: key = AQAcsq9pPGAwFBAAgcvATnOBJTLHVSzCyxsH0g== 2026-03-10T05:54:36.493 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T05:54:36.494 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/ceph/ceph.client.1.keyring 2026-03-10T05:54:36.494 DEBUG:teuthology.orchestra.run.vm09:> sudo chmod 0644 /etc/ceph/ceph.client.1.keyring 2026-03-10T05:54:36.527 INFO:tasks.ceph:Waiting until ceph daemons up and pgs clean... 2026-03-10T05:54:36.527 INFO:tasks.cephadm.ceph_manager.ceph:waiting for mgr available 2026-03-10T05:54:36.527 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph mgr dump --format=json 2026-03-10T05:54:36.686 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:36.802 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:36 vm01 ceph-mon[47299]: pgmap v44: 1 pgs: 1 unknown; 0 B data, 211 MiB used, 160 GiB / 160 GiB avail 2026-03-10T05:54:36.802 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:36 vm01 ceph-mon[47299]: from='client.? 192.168.123.109:0/3422520165' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T05:54:36.802 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:36 vm01 ceph-mon[47299]: from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T05:54:36.802 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:36 vm01 ceph-mon[47299]: from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-10T05:54:36.926 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:54:37.094 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":18,"flags":0,"active_gid":14217,"active_name":"vm01.wforik","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6800","nonce":4199812616},{"type":"v1","addr":"192.168.123.101:6801","nonce":4199812616}]},"active_addr":"192.168.123.101:6801/4199812616","active_change":"2026-03-10T05:53:35.455620+0000","active_mgr_features":4540701547738038271,"available":true,"standbys":[{"gid":14240,"name":"vm09.uxbryv","mgr_features":4540701547738038271,"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}]}],"modules":["cephadm","dashboard","iostat","nfs","prometheus","restful"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{"dashboard":"https://192.168.123.101:8443/","prometheus":"http://192.168.123.101:9283/"},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"reef":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"squid":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"]},"force_disabled_modules":{},"last_failure_osd_epoch":5,"active_clients":[{"name":"libcephsqlite","addrvec":[{"type":"v2","addr":"192.168.123.101:0","nonce":1078677439}]},{"name":"rbd_support","addrvec":[{"type":"v2","addr":"192.168.123.101:0","nonce":289801043}]},{"name":"volumes","addrvec":[{"type":"v2","addr":"192.168.123.101:0","nonce":1309997126}]}]} 2026-03-10T05:54:37.096 INFO:tasks.cephadm.ceph_manager.ceph:mgr available! 2026-03-10T05:54:37.096 INFO:tasks.cephadm.ceph_manager.ceph:waiting for all up 2026-03-10T05:54:37.096 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd dump --format=json 2026-03-10T05:54:37.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:36 vm09 ceph-mon[53707]: pgmap v44: 1 pgs: 1 unknown; 0 B data, 211 MiB used, 160 GiB / 160 GiB avail 2026-03-10T05:54:37.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:36 vm09 ceph-mon[53707]: from='client.? 192.168.123.109:0/3422520165' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T05:54:37.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:36 vm09 ceph-mon[53707]: from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T05:54:37.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:36 vm09 ceph-mon[53707]: from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-10T05:54:37.257 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:37.473 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:54:37.473 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":26,"fsid":"100dc3f2-1c45-11f1-8010-b96ca19c2df4","created":"2026-03-10T05:52:41.180004+0000","modified":"2026-03-10T05:54:33.732036+0000","last_up_change":"2026-03-10T05:54:32.729144+0000","last_in_change":"2026-03-10T05:54:12.101073+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":12,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-10T05:54:27.730315+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"23","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":7.8899998664855957,"score_stable":7.8899998664855957,"optimal_score":0.37999999523162842,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"a48f3e69-2204-46a2-aa6e-22512319163b","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6800","nonce":3601312935},{"type":"v1","addr":"192.168.123.109:6801","nonce":3601312935}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6802","nonce":3601312935},{"type":"v1","addr":"192.168.123.109:6803","nonce":3601312935}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6806","nonce":3601312935},{"type":"v1","addr":"192.168.123.109:6807","nonce":3601312935}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6804","nonce":3601312935},{"type":"v1","addr":"192.168.123.109:6805","nonce":3601312935}]},"public_addr":"192.168.123.109:6801/3601312935","cluster_addr":"192.168.123.109:6803/3601312935","heartbeat_back_addr":"192.168.123.109:6807/3601312935","heartbeat_front_addr":"192.168.123.109:6805/3601312935","state":["exists","up"]},{"osd":1,"uuid":"85e73c81-a2da-499a-8cf1-5da75fea2dd5","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6802","nonce":2835686876},{"type":"v1","addr":"192.168.123.101:6803","nonce":2835686876}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6804","nonce":2835686876},{"type":"v1","addr":"192.168.123.101:6805","nonce":2835686876}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6808","nonce":2835686876},{"type":"v1","addr":"192.168.123.101:6809","nonce":2835686876}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6806","nonce":2835686876},{"type":"v1","addr":"192.168.123.101:6807","nonce":2835686876}]},"public_addr":"192.168.123.101:6803/2835686876","cluster_addr":"192.168.123.101:6805/2835686876","heartbeat_back_addr":"192.168.123.101:6809/2835686876","heartbeat_front_addr":"192.168.123.101:6807/2835686876","state":["exists","up"]},{"osd":2,"uuid":"441039ef-c8de-4d0a-aa8d-1950348d8d54","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":19,"up_thru":21,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6808","nonce":2232889601},{"type":"v1","addr":"192.168.123.109:6809","nonce":2232889601}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6810","nonce":2232889601},{"type":"v1","addr":"192.168.123.109:6811","nonce":2232889601}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6814","nonce":2232889601},{"type":"v1","addr":"192.168.123.109:6815","nonce":2232889601}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6812","nonce":2232889601},{"type":"v1","addr":"192.168.123.109:6813","nonce":2232889601}]},"public_addr":"192.168.123.109:6809/2232889601","cluster_addr":"192.168.123.109:6811/2232889601","heartbeat_back_addr":"192.168.123.109:6815/2232889601","heartbeat_front_addr":"192.168.123.109:6813/2232889601","state":["exists","up"]},{"osd":3,"uuid":"2cf13596-358c-4344-9953-0cc5ca8e025f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":20,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6810","nonce":2800357583},{"type":"v1","addr":"192.168.123.101:6811","nonce":2800357583}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6812","nonce":2800357583},{"type":"v1","addr":"192.168.123.101:6813","nonce":2800357583}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6816","nonce":2800357583},{"type":"v1","addr":"192.168.123.101:6817","nonce":2800357583}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6814","nonce":2800357583},{"type":"v1","addr":"192.168.123.101:6815","nonce":2800357583}]},"public_addr":"192.168.123.101:6811/2800357583","cluster_addr":"192.168.123.101:6813/2800357583","heartbeat_back_addr":"192.168.123.101:6817/2800357583","heartbeat_front_addr":"192.168.123.101:6815/2800357583","state":["exists","up"]},{"osd":4,"uuid":"b524239b-feb0-401f-bbd9-f924a65ba45c","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":20,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6816","nonce":1566157596},{"type":"v1","addr":"192.168.123.109:6817","nonce":1566157596}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6818","nonce":1566157596},{"type":"v1","addr":"192.168.123.109:6819","nonce":1566157596}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6822","nonce":1566157596},{"type":"v1","addr":"192.168.123.109:6823","nonce":1566157596}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6820","nonce":1566157596},{"type":"v1","addr":"192.168.123.109:6821","nonce":1566157596}]},"public_addr":"192.168.123.109:6817/1566157596","cluster_addr":"192.168.123.109:6819/1566157596","heartbeat_back_addr":"192.168.123.109:6823/1566157596","heartbeat_front_addr":"192.168.123.109:6821/1566157596","state":["exists","up"]},{"osd":5,"uuid":"11ed13ca-a6e4-48db-9586-874dcc754728","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":21,"up_thru":24,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6818","nonce":1562465694},{"type":"v1","addr":"192.168.123.101:6819","nonce":1562465694}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6820","nonce":1562465694},{"type":"v1","addr":"192.168.123.101:6821","nonce":1562465694}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6824","nonce":1562465694},{"type":"v1","addr":"192.168.123.101:6825","nonce":1562465694}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6822","nonce":1562465694},{"type":"v1","addr":"192.168.123.101:6823","nonce":1562465694}]},"public_addr":"192.168.123.101:6819/1562465694","cluster_addr":"192.168.123.101:6821/1562465694","heartbeat_back_addr":"192.168.123.101:6825/1562465694","heartbeat_front_addr":"192.168.123.101:6823/1562465694","state":["exists","up"]},{"osd":6,"uuid":"f5910813-2e44-4da2-9c8e-48450f4a8fab","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":24,"up_thru":25,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6824","nonce":1216154444},{"type":"v1","addr":"192.168.123.109:6825","nonce":1216154444}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6826","nonce":1216154444},{"type":"v1","addr":"192.168.123.109:6827","nonce":1216154444}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6830","nonce":1216154444},{"type":"v1","addr":"192.168.123.109:6831","nonce":1216154444}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6828","nonce":1216154444},{"type":"v1","addr":"192.168.123.109:6829","nonce":1216154444}]},"public_addr":"192.168.123.109:6825/1216154444","cluster_addr":"192.168.123.109:6827/1216154444","heartbeat_back_addr":"192.168.123.109:6831/1216154444","heartbeat_front_addr":"192.168.123.109:6829/1216154444","state":["exists","up"]},{"osd":7,"uuid":"f4cbd2d8-a341-4674-88e2-75527b726469","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":25,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6826","nonce":1172902724},{"type":"v1","addr":"192.168.123.101:6827","nonce":1172902724}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6828","nonce":1172902724},{"type":"v1","addr":"192.168.123.101:6829","nonce":1172902724}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6832","nonce":1172902724},{"type":"v1","addr":"192.168.123.101:6833","nonce":1172902724}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6830","nonce":1172902724},{"type":"v1","addr":"192.168.123.101:6831","nonce":1172902724}]},"public_addr":"192.168.123.101:6827/1172902724","cluster_addr":"192.168.123.101:6829/1172902724","heartbeat_back_addr":"192.168.123.101:6833/1172902724","heartbeat_front_addr":"192.168.123.101:6831/1172902724","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:54:21.619546+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:54:21.765397+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:54:24.139079+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:54:24.043381+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:54:26.722231+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:54:27.434731+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:54:29.836722+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:54:31.300808+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.101:6801/1561074913":"2026-03-11T05:53:35.455537+0000","192.168.123.101:0/877175877":"2026-03-11T05:53:35.455537+0000","192.168.123.101:0/100373818":"2026-03-11T05:53:35.455537+0000","192.168.123.101:0/3534784264":"2026-03-11T05:53:05.831349+0000","192.168.123.101:6800/1229025911":"2026-03-11T05:52:53.419293+0000","192.168.123.101:6800/1371156494":"2026-03-11T05:53:05.831349+0000","192.168.123.101:0/500832301":"2026-03-11T05:52:53.419293+0000","192.168.123.101:6801/1229025911":"2026-03-11T05:52:53.419293+0000","192.168.123.101:6801/1371156494":"2026-03-11T05:53:05.831349+0000","192.168.123.101:0/3827013041":"2026-03-11T05:52:53.419293+0000","192.168.123.101:0/2988507782":"2026-03-11T05:52:53.419293+0000","192.168.123.101:0/1403556378":"2026-03-11T05:53:05.831349+0000","192.168.123.101:0/2020700462":"2026-03-11T05:53:05.831349+0000","192.168.123.101:6800/1561074913":"2026-03-11T05:53:35.455537+0000","192.168.123.101:0/1369389137":"2026-03-11T05:53:35.455537+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-10T05:54:37.639 INFO:tasks.cephadm.ceph_manager.ceph:all up! 2026-03-10T05:54:37.640 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd dump --format=json 2026-03-10T05:54:37.808 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:37.866 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:37 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/2890191744' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-10T05:54:37.866 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:37 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/35576220' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:54:38.038 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:54:38.038 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":26,"fsid":"100dc3f2-1c45-11f1-8010-b96ca19c2df4","created":"2026-03-10T05:52:41.180004+0000","modified":"2026-03-10T05:54:33.732036+0000","last_up_change":"2026-03-10T05:54:32.729144+0000","last_in_change":"2026-03-10T05:54:12.101073+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":12,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-10T05:54:27.730315+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"23","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":7.8899998664855957,"score_stable":7.8899998664855957,"optimal_score":0.37999999523162842,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"a48f3e69-2204-46a2-aa6e-22512319163b","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6800","nonce":3601312935},{"type":"v1","addr":"192.168.123.109:6801","nonce":3601312935}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6802","nonce":3601312935},{"type":"v1","addr":"192.168.123.109:6803","nonce":3601312935}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6806","nonce":3601312935},{"type":"v1","addr":"192.168.123.109:6807","nonce":3601312935}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6804","nonce":3601312935},{"type":"v1","addr":"192.168.123.109:6805","nonce":3601312935}]},"public_addr":"192.168.123.109:6801/3601312935","cluster_addr":"192.168.123.109:6803/3601312935","heartbeat_back_addr":"192.168.123.109:6807/3601312935","heartbeat_front_addr":"192.168.123.109:6805/3601312935","state":["exists","up"]},{"osd":1,"uuid":"85e73c81-a2da-499a-8cf1-5da75fea2dd5","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6802","nonce":2835686876},{"type":"v1","addr":"192.168.123.101:6803","nonce":2835686876}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6804","nonce":2835686876},{"type":"v1","addr":"192.168.123.101:6805","nonce":2835686876}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6808","nonce":2835686876},{"type":"v1","addr":"192.168.123.101:6809","nonce":2835686876}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6806","nonce":2835686876},{"type":"v1","addr":"192.168.123.101:6807","nonce":2835686876}]},"public_addr":"192.168.123.101:6803/2835686876","cluster_addr":"192.168.123.101:6805/2835686876","heartbeat_back_addr":"192.168.123.101:6809/2835686876","heartbeat_front_addr":"192.168.123.101:6807/2835686876","state":["exists","up"]},{"osd":2,"uuid":"441039ef-c8de-4d0a-aa8d-1950348d8d54","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":19,"up_thru":21,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6808","nonce":2232889601},{"type":"v1","addr":"192.168.123.109:6809","nonce":2232889601}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6810","nonce":2232889601},{"type":"v1","addr":"192.168.123.109:6811","nonce":2232889601}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6814","nonce":2232889601},{"type":"v1","addr":"192.168.123.109:6815","nonce":2232889601}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6812","nonce":2232889601},{"type":"v1","addr":"192.168.123.109:6813","nonce":2232889601}]},"public_addr":"192.168.123.109:6809/2232889601","cluster_addr":"192.168.123.109:6811/2232889601","heartbeat_back_addr":"192.168.123.109:6815/2232889601","heartbeat_front_addr":"192.168.123.109:6813/2232889601","state":["exists","up"]},{"osd":3,"uuid":"2cf13596-358c-4344-9953-0cc5ca8e025f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":20,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6810","nonce":2800357583},{"type":"v1","addr":"192.168.123.101:6811","nonce":2800357583}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6812","nonce":2800357583},{"type":"v1","addr":"192.168.123.101:6813","nonce":2800357583}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6816","nonce":2800357583},{"type":"v1","addr":"192.168.123.101:6817","nonce":2800357583}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6814","nonce":2800357583},{"type":"v1","addr":"192.168.123.101:6815","nonce":2800357583}]},"public_addr":"192.168.123.101:6811/2800357583","cluster_addr":"192.168.123.101:6813/2800357583","heartbeat_back_addr":"192.168.123.101:6817/2800357583","heartbeat_front_addr":"192.168.123.101:6815/2800357583","state":["exists","up"]},{"osd":4,"uuid":"b524239b-feb0-401f-bbd9-f924a65ba45c","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":20,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6816","nonce":1566157596},{"type":"v1","addr":"192.168.123.109:6817","nonce":1566157596}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6818","nonce":1566157596},{"type":"v1","addr":"192.168.123.109:6819","nonce":1566157596}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6822","nonce":1566157596},{"type":"v1","addr":"192.168.123.109:6823","nonce":1566157596}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6820","nonce":1566157596},{"type":"v1","addr":"192.168.123.109:6821","nonce":1566157596}]},"public_addr":"192.168.123.109:6817/1566157596","cluster_addr":"192.168.123.109:6819/1566157596","heartbeat_back_addr":"192.168.123.109:6823/1566157596","heartbeat_front_addr":"192.168.123.109:6821/1566157596","state":["exists","up"]},{"osd":5,"uuid":"11ed13ca-a6e4-48db-9586-874dcc754728","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":21,"up_thru":24,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6818","nonce":1562465694},{"type":"v1","addr":"192.168.123.101:6819","nonce":1562465694}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6820","nonce":1562465694},{"type":"v1","addr":"192.168.123.101:6821","nonce":1562465694}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6824","nonce":1562465694},{"type":"v1","addr":"192.168.123.101:6825","nonce":1562465694}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6822","nonce":1562465694},{"type":"v1","addr":"192.168.123.101:6823","nonce":1562465694}]},"public_addr":"192.168.123.101:6819/1562465694","cluster_addr":"192.168.123.101:6821/1562465694","heartbeat_back_addr":"192.168.123.101:6825/1562465694","heartbeat_front_addr":"192.168.123.101:6823/1562465694","state":["exists","up"]},{"osd":6,"uuid":"f5910813-2e44-4da2-9c8e-48450f4a8fab","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":24,"up_thru":25,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6824","nonce":1216154444},{"type":"v1","addr":"192.168.123.109:6825","nonce":1216154444}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6826","nonce":1216154444},{"type":"v1","addr":"192.168.123.109:6827","nonce":1216154444}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6830","nonce":1216154444},{"type":"v1","addr":"192.168.123.109:6831","nonce":1216154444}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6828","nonce":1216154444},{"type":"v1","addr":"192.168.123.109:6829","nonce":1216154444}]},"public_addr":"192.168.123.109:6825/1216154444","cluster_addr":"192.168.123.109:6827/1216154444","heartbeat_back_addr":"192.168.123.109:6831/1216154444","heartbeat_front_addr":"192.168.123.109:6829/1216154444","state":["exists","up"]},{"osd":7,"uuid":"f4cbd2d8-a341-4674-88e2-75527b726469","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":25,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6826","nonce":1172902724},{"type":"v1","addr":"192.168.123.101:6827","nonce":1172902724}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6828","nonce":1172902724},{"type":"v1","addr":"192.168.123.101:6829","nonce":1172902724}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6832","nonce":1172902724},{"type":"v1","addr":"192.168.123.101:6833","nonce":1172902724}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6830","nonce":1172902724},{"type":"v1","addr":"192.168.123.101:6831","nonce":1172902724}]},"public_addr":"192.168.123.101:6827/1172902724","cluster_addr":"192.168.123.101:6829/1172902724","heartbeat_back_addr":"192.168.123.101:6833/1172902724","heartbeat_front_addr":"192.168.123.101:6831/1172902724","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:54:21.619546+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:54:21.765397+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:54:24.139079+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:54:24.043381+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:54:26.722231+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:54:27.434731+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:54:29.836722+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:54:31.300808+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.101:6801/1561074913":"2026-03-11T05:53:35.455537+0000","192.168.123.101:0/877175877":"2026-03-11T05:53:35.455537+0000","192.168.123.101:0/100373818":"2026-03-11T05:53:35.455537+0000","192.168.123.101:0/3534784264":"2026-03-11T05:53:05.831349+0000","192.168.123.101:6800/1229025911":"2026-03-11T05:52:53.419293+0000","192.168.123.101:6800/1371156494":"2026-03-11T05:53:05.831349+0000","192.168.123.101:0/500832301":"2026-03-11T05:52:53.419293+0000","192.168.123.101:6801/1229025911":"2026-03-11T05:52:53.419293+0000","192.168.123.101:6801/1371156494":"2026-03-11T05:53:05.831349+0000","192.168.123.101:0/3827013041":"2026-03-11T05:52:53.419293+0000","192.168.123.101:0/2988507782":"2026-03-11T05:52:53.419293+0000","192.168.123.101:0/1403556378":"2026-03-11T05:53:05.831349+0000","192.168.123.101:0/2020700462":"2026-03-11T05:53:05.831349+0000","192.168.123.101:6800/1561074913":"2026-03-11T05:53:35.455537+0000","192.168.123.101:0/1369389137":"2026-03-11T05:53:35.455537+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-10T05:54:38.188 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph tell osd.0 flush_pg_stats 2026-03-10T05:54:38.188 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph tell osd.1 flush_pg_stats 2026-03-10T05:54:38.188 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph tell osd.2 flush_pg_stats 2026-03-10T05:54:38.188 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph tell osd.3 flush_pg_stats 2026-03-10T05:54:38.188 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph tell osd.4 flush_pg_stats 2026-03-10T05:54:38.188 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph tell osd.5 flush_pg_stats 2026-03-10T05:54:38.188 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph tell osd.6 flush_pg_stats 2026-03-10T05:54:38.188 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph tell osd.7 flush_pg_stats 2026-03-10T05:54:38.209 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:37 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/2890191744' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-10T05:54:38.209 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:37 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/35576220' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:54:38.844 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:38.859 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:38.973 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:38.986 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:38.998 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:39.036 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:38 vm01 ceph-mon[47299]: pgmap v45: 1 pgs: 1 active+clean; 0 B data, 212 MiB used, 160 GiB / 160 GiB avail 2026-03-10T05:54:39.036 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:38 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/1015714857' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:54:39.200 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:39.209 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:38 vm09 ceph-mon[53707]: pgmap v45: 1 pgs: 1 active+clean; 0 B data, 212 MiB used, 160 GiB / 160 GiB avail 2026-03-10T05:54:39.209 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:38 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/1015714857' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:54:39.338 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:39.343 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:39.484 INFO:teuthology.orchestra.run.vm01.stdout:73014444036 2026-03-10T05:54:39.485 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd last-stat-seq osd.1 2026-03-10T05:54:39.560 INFO:teuthology.orchestra.run.vm01.stdout:73014444036 2026-03-10T05:54:39.560 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd last-stat-seq osd.0 2026-03-10T05:54:39.909 INFO:teuthology.orchestra.run.vm01.stdout:103079215107 2026-03-10T05:54:39.910 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd last-stat-seq osd.6 2026-03-10T05:54:39.989 INFO:teuthology.orchestra.run.vm01.stdout:85899345924 2026-03-10T05:54:39.989 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd last-stat-seq osd.3 2026-03-10T05:54:40.082 INFO:teuthology.orchestra.run.vm01.stdout:81604378628 2026-03-10T05:54:40.083 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd last-stat-seq osd.2 2026-03-10T05:54:40.135 INFO:teuthology.orchestra.run.vm01.stdout:85899345924 2026-03-10T05:54:40.135 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd last-stat-seq osd.4 2026-03-10T05:54:40.182 INFO:teuthology.orchestra.run.vm01.stdout:107374182403 2026-03-10T05:54:40.182 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd last-stat-seq osd.7 2026-03-10T05:54:40.218 INFO:teuthology.orchestra.run.vm01.stdout:90194313220 2026-03-10T05:54:40.218 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd last-stat-seq osd.5 2026-03-10T05:54:40.240 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:40.267 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:40.637 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:40.676 INFO:teuthology.orchestra.run.vm01.stdout:73014444036 2026-03-10T05:54:40.799 INFO:teuthology.orchestra.run.vm01.stdout:73014444036 2026-03-10T05:54:40.931 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:40 vm01 ceph-mon[47299]: pgmap v46: 1 pgs: 1 active+clean; 0 B data, 212 MiB used, 160 GiB / 160 GiB avail 2026-03-10T05:54:40.931 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:40 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/2865021158' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T05:54:40.932 INFO:tasks.cephadm.ceph_manager.ceph:need seq 73014444036 got 73014444036 for osd.1 2026-03-10T05:54:40.932 DEBUG:teuthology.parallel:result is None 2026-03-10T05:54:41.058 INFO:teuthology.orchestra.run.vm01.stdout:103079215107 2026-03-10T05:54:41.093 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:41.132 INFO:tasks.cephadm.ceph_manager.ceph:need seq 73014444036 got 73014444036 for osd.0 2026-03-10T05:54:41.132 DEBUG:teuthology.parallel:result is None 2026-03-10T05:54:41.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:40 vm09 ceph-mon[53707]: pgmap v46: 1 pgs: 1 active+clean; 0 B data, 212 MiB used, 160 GiB / 160 GiB avail 2026-03-10T05:54:41.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:40 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/2865021158' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T05:54:41.278 INFO:tasks.cephadm.ceph_manager.ceph:need seq 103079215107 got 103079215107 for osd.6 2026-03-10T05:54:41.278 DEBUG:teuthology.parallel:result is None 2026-03-10T05:54:41.359 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:41.367 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:41.368 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:41.384 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:41.406 INFO:teuthology.orchestra.run.vm01.stdout:85899345923 2026-03-10T05:54:41.621 INFO:tasks.cephadm.ceph_manager.ceph:need seq 85899345924 got 85899345923 for osd.3 2026-03-10T05:54:41.859 INFO:teuthology.orchestra.run.vm01.stdout:85899345924 2026-03-10T05:54:41.882 INFO:teuthology.orchestra.run.vm01.stdout:81604378628 2026-03-10T05:54:41.901 INFO:teuthology.orchestra.run.vm01.stdout:90194313220 2026-03-10T05:54:41.916 INFO:teuthology.orchestra.run.vm01.stdout:107374182403 2026-03-10T05:54:42.047 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:41 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/946319912' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T05:54:42.047 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:41 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/2684190281' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-10T05:54:42.047 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:41 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/3052607852' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-10T05:54:42.086 INFO:tasks.cephadm.ceph_manager.ceph:need seq 107374182403 got 107374182403 for osd.7 2026-03-10T05:54:42.086 DEBUG:teuthology.parallel:result is None 2026-03-10T05:54:42.119 INFO:tasks.cephadm.ceph_manager.ceph:need seq 81604378628 got 81604378628 for osd.2 2026-03-10T05:54:42.119 DEBUG:teuthology.parallel:result is None 2026-03-10T05:54:42.125 INFO:tasks.cephadm.ceph_manager.ceph:need seq 90194313220 got 90194313220 for osd.5 2026-03-10T05:54:42.125 DEBUG:teuthology.parallel:result is None 2026-03-10T05:54:42.126 INFO:tasks.cephadm.ceph_manager.ceph:need seq 85899345924 got 85899345924 for osd.4 2026-03-10T05:54:42.126 DEBUG:teuthology.parallel:result is None 2026-03-10T05:54:42.209 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:41 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/946319912' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T05:54:42.209 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:41 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/2684190281' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-10T05:54:42.209 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:41 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/3052607852' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-10T05:54:42.622 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph osd last-stat-seq osd.3 2026-03-10T05:54:42.800 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:42.921 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:42 vm01 ceph-mon[47299]: pgmap v47: 1 pgs: 1 active+clean; 0 B data, 212 MiB used, 160 GiB / 160 GiB avail 2026-03-10T05:54:42.921 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:42 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/2758434792' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-10T05:54:42.921 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:42 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/1340929011' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-10T05:54:42.921 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:42 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/3369239302' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-10T05:54:42.921 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:42 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/2301930506' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-10T05:54:43.040 INFO:teuthology.orchestra.run.vm01.stdout:85899345924 2026-03-10T05:54:43.209 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:42 vm09 ceph-mon[53707]: pgmap v47: 1 pgs: 1 active+clean; 0 B data, 212 MiB used, 160 GiB / 160 GiB avail 2026-03-10T05:54:43.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:42 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/2758434792' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-10T05:54:43.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:42 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/1340929011' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-10T05:54:43.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:42 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/3369239302' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-10T05:54:43.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:42 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/2301930506' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-10T05:54:43.212 INFO:tasks.cephadm.ceph_manager.ceph:need seq 85899345924 got 85899345924 for osd.3 2026-03-10T05:54:43.212 DEBUG:teuthology.parallel:result is None 2026-03-10T05:54:43.212 INFO:tasks.cephadm.ceph_manager.ceph:waiting for clean 2026-03-10T05:54:43.213 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph pg dump --format=json 2026-03-10T05:54:43.392 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:43.621 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:54:43.622 INFO:teuthology.orchestra.run.vm01.stderr:dumped all 2026-03-10T05:54:43.769 INFO:teuthology.orchestra.run.vm01.stdout:{"pg_ready":true,"pg_map":{"version":47,"stamp":"2026-03-10T05:54:41.701929+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":9,"num_read_kb":5,"num_write":11,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":1,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":3,"ondisk_log_size":3,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":3,"num_osds":8,"num_per_pool_osds":8,"num_per_pool_omap_osds":3,"kb":167739392,"kb_used":216688,"kb_used_data":1952,"kb_used_omap":12,"kb_used_meta":214515,"kb_avail":167522704,"statfs":{"total":171765137408,"available":171543248896,"internally_reserved":0,"allocated":1998848,"data_stored":748960,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":12712,"internal_metadata":219663960},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"4.000514"},"pg_stats":[{"pgid":"1.0","version":"22'3","reported_seq":15,"reported_epoch":26,"state":"active+clean","last_fresh":"2026-03-10T05:54:33.944345+0000","last_change":"2026-03-10T05:54:33.943426+0000","last_active":"2026-03-10T05:54:33.944345+0000","last_peered":"2026-03-10T05:54:33.944345+0000","last_clean":"2026-03-10T05:54:33.944345+0000","last_became_active":"2026-03-10T05:54:33.740389+0000","last_became_peered":"2026-03-10T05:54:33.740389+0000","last_unstale":"2026-03-10T05:54:33.944345+0000","last_undegraded":"2026-03-10T05:54:33.944345+0000","last_fullsized":"2026-03-10T05:54:33.944345+0000","mapping_epoch":25,"log_start":"0'0","ondisk_log_start":"0'0","created":21,"last_epoch_clean":26,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T05:54:28.717222+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T05:54:28.717222+0000","last_clean_scrub_stamp":"2026-03-10T05:54:28.717222+0000","objects_scrubbed":0,"log_size":3,"log_dups_size":0,"ondisk_log_size":3,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T08:57:20.670811+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":9,"num_read_kb":5,"num_write":11,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":1,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[6,5,2],"acting":[6,5,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":6,"acting_primary":6,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":9,"num_read_kb":5,"num_write":11,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":1,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":3,"ondisk_log_size":3,"up":3,"acting":3,"num_store_stats":3}],"osd_stats":[{"osd":7,"up_from":25,"seq":107374182403,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27084,"kb_used_data":244,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940340,"statfs":{"total":21470642176,"available":21442908160,"internally_reserved":0,"allocated":249856,"data_stored":93620,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1585,"internal_metadata":27457999},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":6,"up_from":24,"seq":103079215107,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27084,"kb_used_data":244,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940340,"statfs":{"total":21470642176,"available":21442908160,"internally_reserved":0,"allocated":249856,"data_stored":93620,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1589,"internal_metadata":27457995},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":5,"up_from":21,"seq":90194313220,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27084,"kb_used_data":244,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940340,"statfs":{"total":21470642176,"available":21442908160,"internally_reserved":0,"allocated":249856,"data_stored":93620,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":20,"seq":85899345924,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27092,"kb_used_data":244,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940332,"statfs":{"total":21470642176,"available":21442899968,"internally_reserved":0,"allocated":249856,"data_stored":93620,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":20,"seq":85899345924,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27092,"kb_used_data":244,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940332,"statfs":{"total":21470642176,"available":21442899968,"internally_reserved":0,"allocated":249856,"data_stored":93620,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1588,"internal_metadata":27457996},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":19,"seq":81604378628,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27080,"kb_used_data":244,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940344,"statfs":{"total":21470642176,"available":21442912256,"internally_reserved":0,"allocated":249856,"data_stored":93620,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":17,"seq":73014444037,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27092,"kb_used_data":244,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940332,"statfs":{"total":21470642176,"available":21442899968,"internally_reserved":0,"allocated":249856,"data_stored":93620,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":17,"seq":73014444037,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27080,"kb_used_data":244,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940344,"statfs":{"total":21470642176,"available":21442912256,"internally_reserved":0,"allocated":249856,"data_stored":93620,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-10T05:54:43.770 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph pg dump --format=json 2026-03-10T05:54:43.932 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:43.997 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:43 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/423351820' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-10T05:54:43.997 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:43 vm01 ceph-mon[47299]: from='client.14532 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T05:54:44.144 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:54:44.144 INFO:teuthology.orchestra.run.vm01.stderr:dumped all 2026-03-10T05:54:44.209 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:43 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/423351820' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-10T05:54:44.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:43 vm09 ceph-mon[53707]: from='client.14532 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T05:54:44.309 INFO:teuthology.orchestra.run.vm01.stdout:{"pg_ready":true,"pg_map":{"version":48,"stamp":"2026-03-10T05:54:43.702213+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":9,"num_read_kb":5,"num_write":11,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":1,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":3,"ondisk_log_size":3,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":3,"num_osds":8,"num_per_pool_osds":8,"num_per_pool_omap_osds":3,"kb":167739392,"kb_used":216688,"kb_used_data":1952,"kb_used_omap":12,"kb_used_meta":214515,"kb_avail":167522704,"statfs":{"total":171765137408,"available":171543248896,"internally_reserved":0,"allocated":1998848,"data_stored":748960,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":12712,"internal_metadata":219663960},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"6.000798"},"pg_stats":[{"pgid":"1.0","version":"22'3","reported_seq":15,"reported_epoch":26,"state":"active+clean","last_fresh":"2026-03-10T05:54:33.944345+0000","last_change":"2026-03-10T05:54:33.943426+0000","last_active":"2026-03-10T05:54:33.944345+0000","last_peered":"2026-03-10T05:54:33.944345+0000","last_clean":"2026-03-10T05:54:33.944345+0000","last_became_active":"2026-03-10T05:54:33.740389+0000","last_became_peered":"2026-03-10T05:54:33.740389+0000","last_unstale":"2026-03-10T05:54:33.944345+0000","last_undegraded":"2026-03-10T05:54:33.944345+0000","last_fullsized":"2026-03-10T05:54:33.944345+0000","mapping_epoch":25,"log_start":"0'0","ondisk_log_start":"0'0","created":21,"last_epoch_clean":26,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T05:54:28.717222+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T05:54:28.717222+0000","last_clean_scrub_stamp":"2026-03-10T05:54:28.717222+0000","objects_scrubbed":0,"log_size":3,"log_dups_size":0,"ondisk_log_size":3,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T08:57:20.670811+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":9,"num_read_kb":5,"num_write":11,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":1,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[6,5,2],"acting":[6,5,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":6,"acting_primary":6,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":9,"num_read_kb":5,"num_write":11,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":1,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":3,"ondisk_log_size":3,"up":3,"acting":3,"num_store_stats":3}],"osd_stats":[{"osd":7,"up_from":25,"seq":107374182404,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27084,"kb_used_data":244,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940340,"statfs":{"total":21470642176,"available":21442908160,"internally_reserved":0,"allocated":249856,"data_stored":93620,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1585,"internal_metadata":27457999},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":6,"up_from":24,"seq":103079215108,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27084,"kb_used_data":244,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940340,"statfs":{"total":21470642176,"available":21442908160,"internally_reserved":0,"allocated":249856,"data_stored":93620,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1589,"internal_metadata":27457995},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":5,"up_from":21,"seq":90194313220,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27084,"kb_used_data":244,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940340,"statfs":{"total":21470642176,"available":21442908160,"internally_reserved":0,"allocated":249856,"data_stored":93620,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":20,"seq":85899345925,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27092,"kb_used_data":244,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940332,"statfs":{"total":21470642176,"available":21442899968,"internally_reserved":0,"allocated":249856,"data_stored":93620,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":20,"seq":85899345925,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27092,"kb_used_data":244,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940332,"statfs":{"total":21470642176,"available":21442899968,"internally_reserved":0,"allocated":249856,"data_stored":93620,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1588,"internal_metadata":27457996},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":19,"seq":81604378629,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27080,"kb_used_data":244,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940344,"statfs":{"total":21470642176,"available":21442912256,"internally_reserved":0,"allocated":249856,"data_stored":93620,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":17,"seq":73014444037,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27092,"kb_used_data":244,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940332,"statfs":{"total":21470642176,"available":21442899968,"internally_reserved":0,"allocated":249856,"data_stored":93620,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":17,"seq":73014444037,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27080,"kb_used_data":244,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940344,"statfs":{"total":21470642176,"available":21442912256,"internally_reserved":0,"allocated":249856,"data_stored":93620,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-10T05:54:44.309 INFO:tasks.cephadm.ceph_manager.ceph:clean! 2026-03-10T05:54:44.309 INFO:tasks.ceph:Waiting until ceph cluster ceph is healthy... 2026-03-10T05:54:44.309 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy 2026-03-10T05:54:44.309 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph health --format=json 2026-03-10T05:54:44.471 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:44.705 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:54:44.705 INFO:teuthology.orchestra.run.vm01.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-03-10T05:54:44.874 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy done 2026-03-10T05:54:44.874 INFO:tasks.cephadm:Setup complete, yielding 2026-03-10T05:54:44.874 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-10T05:54:44.876 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm01.local 2026-03-10T05:54:44.876 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- bash -c 'ceph orch status' 2026-03-10T05:54:45.035 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:45.059 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:44 vm01 ceph-mon[47299]: pgmap v48: 1 pgs: 1 active+clean; 0 B data, 212 MiB used, 160 GiB / 160 GiB avail 2026-03-10T05:54:45.059 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:44 vm01 ceph-mon[47299]: from='client.14536 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T05:54:45.059 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:44 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/4294912270' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-10T05:54:45.209 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:44 vm09 ceph-mon[53707]: pgmap v48: 1 pgs: 1 active+clean; 0 B data, 212 MiB used, 160 GiB / 160 GiB avail 2026-03-10T05:54:45.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:44 vm09 ceph-mon[53707]: from='client.14536 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T05:54:45.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:44 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/4294912270' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-10T05:54:45.263 INFO:teuthology.orchestra.run.vm01.stdout:Backend: cephadm 2026-03-10T05:54:45.263 INFO:teuthology.orchestra.run.vm01.stdout:Available: Yes 2026-03-10T05:54:45.263 INFO:teuthology.orchestra.run.vm01.stdout:Paused: No 2026-03-10T05:54:45.411 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- bash -c 'ceph orch ps' 2026-03-10T05:54:45.576 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:45.822 INFO:teuthology.orchestra.run.vm01.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-10T05:54:45.822 INFO:teuthology.orchestra.run.vm01.stdout:alertmanager.vm01 vm01 *:9093,9094 running (51s) 15s ago 86s 22.6M - 0.25.0 c8568f914cd2 639e5539aa67 2026-03-10T05:54:45.822 INFO:teuthology.orchestra.run.vm01.stdout:ceph-exporter.vm01 vm01 *:9926 running (93s) 15s ago 93s 8636k - 19.2.3-678-ge911bdeb 654f31e6858e 4d03abd25a35 2026-03-10T05:54:45.822 INFO:teuthology.orchestra.run.vm01.stdout:ceph-exporter.vm09 vm09 *:9926 running (65s) 15s ago 65s 6555k - 19.2.3-678-ge911bdeb 654f31e6858e 685feb193756 2026-03-10T05:54:45.822 INFO:teuthology.orchestra.run.vm01.stdout:crash.vm01 vm01 running (92s) 15s ago 92s 7654k - 19.2.3-678-ge911bdeb 654f31e6858e 1d73eef3d498 2026-03-10T05:54:45.822 INFO:teuthology.orchestra.run.vm01.stdout:crash.vm09 vm09 running (65s) 15s ago 64s 7654k - 19.2.3-678-ge911bdeb 654f31e6858e dc6057be483c 2026-03-10T05:54:45.822 INFO:teuthology.orchestra.run.vm01.stdout:grafana.vm01 vm01 *:3000 running (49s) 15s ago 79s 82.5M - 10.4.0 c8b91775d855 792467d3a04f 2026-03-10T05:54:45.822 INFO:teuthology.orchestra.run.vm01.stdout:mgr.vm01.wforik vm01 *:9283,8765,8443 running (2m) 15s ago 2m 544M - 19.2.3-678-ge911bdeb 654f31e6858e fd848140b65b 2026-03-10T05:54:45.822 INFO:teuthology.orchestra.run.vm01.stdout:mgr.vm09.uxbryv vm09 *:8443,9283,8765 running (61s) 15s ago 61s 490M - 19.2.3-678-ge911bdeb 654f31e6858e 7d2c10ecee80 2026-03-10T05:54:45.822 INFO:teuthology.orchestra.run.vm01.stdout:mon.vm01 vm01 running (2m) 15s ago 2m 48.2M 2048M 19.2.3-678-ge911bdeb 654f31e6858e d0fd780adbec 2026-03-10T05:54:45.822 INFO:teuthology.orchestra.run.vm01.stdout:mon.vm09 vm09 running (59s) 15s ago 59s 43.5M 2048M 19.2.3-678-ge911bdeb 654f31e6858e 98ffb94408d7 2026-03-10T05:54:45.822 INFO:teuthology.orchestra.run.vm01.stdout:node-exporter.vm01 vm01 *:9100 running (89s) 15s ago 89s 9399k - 1.7.0 72c9c2088986 e27f0b466530 2026-03-10T05:54:45.822 INFO:teuthology.orchestra.run.vm01.stdout:node-exporter.vm09 vm09 *:9100 running (62s) 15s ago 62s 9118k - 1.7.0 72c9c2088986 73b4f78f9107 2026-03-10T05:54:45.822 INFO:teuthology.orchestra.run.vm01.stdout:osd.0 vm09 running (26s) 15s ago 26s 52.9M 4096M 19.2.3-678-ge911bdeb 654f31e6858e 8ce2792179a6 2026-03-10T05:54:45.822 INFO:teuthology.orchestra.run.vm01.stdout:osd.1 vm01 running (27s) 15s ago 26s 54.1M 4096M 19.2.3-678-ge911bdeb 654f31e6858e b9ecca506167 2026-03-10T05:54:45.823 INFO:teuthology.orchestra.run.vm01.stdout:osd.2 vm09 running (24s) 15s ago 24s 53.1M 4096M 19.2.3-678-ge911bdeb 654f31e6858e 6cc8199a1748 2026-03-10T05:54:45.823 INFO:teuthology.orchestra.run.vm01.stdout:osd.3 vm01 running (24s) 15s ago 24s 56.3M 4096M 19.2.3-678-ge911bdeb 654f31e6858e 2c0023449970 2026-03-10T05:54:45.823 INFO:teuthology.orchestra.run.vm01.stdout:osd.4 vm09 running (21s) 15s ago 21s 29.9M 4096M 19.2.3-678-ge911bdeb 654f31e6858e c12d4f24588d 2026-03-10T05:54:45.823 INFO:teuthology.orchestra.run.vm01.stdout:osd.5 vm01 running (21s) 15s ago 21s 32.6M 4096M 19.2.3-678-ge911bdeb 654f31e6858e e85aa3b83232 2026-03-10T05:54:45.823 INFO:teuthology.orchestra.run.vm01.stdout:osd.6 vm09 running (19s) 15s ago 18s 26.1M 4096M 19.2.3-678-ge911bdeb 654f31e6858e c32ea635ccf3 2026-03-10T05:54:45.823 INFO:teuthology.orchestra.run.vm01.stdout:osd.7 vm01 running (17s) 15s ago 17s 13.8M 4096M 19.2.3-678-ge911bdeb 654f31e6858e d6d776c1ae0f 2026-03-10T05:54:45.823 INFO:teuthology.orchestra.run.vm01.stdout:prometheus.vm01 vm01 *:9095 running (48s) 15s ago 74s 32.2M - 2.51.0 1d3b7f56885b c505b49771f5 2026-03-10T05:54:45.897 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:45 vm01 ceph-mon[47299]: from='client.14544 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:54:45.979 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- bash -c 'ceph orch ls' 2026-03-10T05:54:46.156 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:46.179 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:45 vm01 sudo[80108]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-10T05:54:46.179 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:45 vm01 sudo[80108]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-10T05:54:46.179 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:45 vm01 sudo[80108]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-10T05:54:46.179 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:45 vm01 sudo[80108]: pam_unix(sudo:session): session closed for user root 2026-03-10T05:54:46.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:45 vm09 ceph-mon[53707]: from='client.14544 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:54:46.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:45 vm09 sudo[67113]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-10T05:54:46.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:45 vm09 sudo[67113]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-10T05:54:46.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:45 vm09 sudo[67113]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-10T05:54:46.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:46 vm09 sudo[67113]: pam_unix(sudo:session): session closed for user root 2026-03-10T05:54:46.390 INFO:teuthology.orchestra.run.vm01.stdout:NAME PORTS RUNNING REFRESHED AGE PLACEMENT 2026-03-10T05:54:46.390 INFO:teuthology.orchestra.run.vm01.stdout:alertmanager ?:9093,9094 1/1 15s ago 105s count:1 2026-03-10T05:54:46.390 INFO:teuthology.orchestra.run.vm01.stdout:ceph-exporter ?:9926 2/2 16s ago 107s * 2026-03-10T05:54:46.390 INFO:teuthology.orchestra.run.vm01.stdout:crash 2/2 16s ago 107s * 2026-03-10T05:54:46.390 INFO:teuthology.orchestra.run.vm01.stdout:grafana ?:3000 1/1 15s ago 106s count:1 2026-03-10T05:54:46.390 INFO:teuthology.orchestra.run.vm01.stdout:mgr 2/2 16s ago 107s count:2 2026-03-10T05:54:46.390 INFO:teuthology.orchestra.run.vm01.stdout:mon 2/2 16s ago 91s vm01:192.168.123.101=vm01;vm09:192.168.123.109=vm09;count:2 2026-03-10T05:54:46.390 INFO:teuthology.orchestra.run.vm01.stdout:node-exporter ?:9100 2/2 16s ago 105s * 2026-03-10T05:54:46.390 INFO:teuthology.orchestra.run.vm01.stdout:osd.all-available-devices 8 16s ago 51s * 2026-03-10T05:54:46.390 INFO:teuthology.orchestra.run.vm01.stdout:prometheus ?:9095 1/1 15s ago 106s count:1 2026-03-10T05:54:46.535 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- bash -c 'ceph orch host ls' 2026-03-10T05:54:46.699 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:46.930 INFO:teuthology.orchestra.run.vm01.stdout:HOST ADDR LABELS STATUS 2026-03-10T05:54:46.931 INFO:teuthology.orchestra.run.vm01.stdout:vm01 192.168.123.101 2026-03-10T05:54:46.931 INFO:teuthology.orchestra.run.vm01.stdout:vm09 192.168.123.109 2026-03-10T05:54:46.931 INFO:teuthology.orchestra.run.vm01.stdout:2 hosts in cluster 2026-03-10T05:54:47.052 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:46 vm01 ceph-mon[47299]: pgmap v49: 1 pgs: 1 active+clean; 0 B data, 212 MiB used, 160 GiB / 160 GiB avail 2026-03-10T05:54:47.053 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:46 vm01 ceph-mon[47299]: from='client.14548 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:54:47.053 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:46 vm01 ceph-mon[47299]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T05:54:47.053 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:46 vm01 ceph-mon[47299]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T05:54:47.053 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:46 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-10T05:54:47.053 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:46 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T05:54:47.053 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:46 vm01 ceph-mon[47299]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T05:54:47.053 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:46 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-10T05:54:47.053 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:46 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T05:54:47.053 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:46 vm01 ceph-mon[47299]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T05:54:47.053 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:46 vm01 ceph-mon[47299]: from='client.14558 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:54:47.080 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- bash -c 'ceph orch device ls' 2026-03-10T05:54:47.209 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:46 vm09 ceph-mon[53707]: pgmap v49: 1 pgs: 1 active+clean; 0 B data, 212 MiB used, 160 GiB / 160 GiB avail 2026-03-10T05:54:47.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:46 vm09 ceph-mon[53707]: from='client.14548 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:54:47.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:46 vm09 ceph-mon[53707]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T05:54:47.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:46 vm09 ceph-mon[53707]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T05:54:47.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-10T05:54:47.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T05:54:47.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:46 vm09 ceph-mon[53707]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T05:54:47.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-10T05:54:47.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:46 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T05:54:47.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:46 vm09 ceph-mon[53707]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T05:54:47.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:46 vm09 ceph-mon[53707]: from='client.14558 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:54:47.243 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:47.488 INFO:teuthology.orchestra.run.vm01.stdout:HOST PATH TYPE DEVICE ID SIZE AVAILABLE REFRESHED REJECT REASONS 2026-03-10T05:54:47.488 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 15s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-10T05:54:47.488 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/vdb hdd DWNBRSTVMM01001 20.0G No 15s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T05:54:47.488 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/vdc hdd DWNBRSTVMM01002 20.0G No 15s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T05:54:47.488 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/vdd hdd DWNBRSTVMM01003 20.0G No 15s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T05:54:47.488 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/vde hdd DWNBRSTVMM01004 20.0G No 15s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T05:54:47.488 INFO:teuthology.orchestra.run.vm01.stdout:vm09 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 16s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-10T05:54:47.488 INFO:teuthology.orchestra.run.vm01.stdout:vm09 /dev/vdb hdd DWNBRSTVMM09001 20.0G No 16s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T05:54:47.488 INFO:teuthology.orchestra.run.vm01.stdout:vm09 /dev/vdc hdd DWNBRSTVMM09002 20.0G No 16s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T05:54:47.488 INFO:teuthology.orchestra.run.vm01.stdout:vm09 /dev/vdd hdd DWNBRSTVMM09003 20.0G No 16s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T05:54:47.488 INFO:teuthology.orchestra.run.vm01.stdout:vm09 /dev/vde hdd DWNBRSTVMM09004 20.0G No 16s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T05:54:47.657 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-10T05:54:47.659 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm01.local 2026-03-10T05:54:47.659 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- bash -c 'ceph orch host label add `hostname` foo' 2026-03-10T05:54:47.817 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:47.889 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:47 vm01 ceph-mon[47299]: from='client.14562 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:54:47.889 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:47 vm01 ceph-mon[47299]: from='client.14566 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:54:48.070 INFO:teuthology.orchestra.run.vm01.stdout:Added label foo to host vm01 2026-03-10T05:54:48.209 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:47 vm09 ceph-mon[53707]: from='client.14562 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:54:48.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:47 vm09 ceph-mon[53707]: from='client.14566 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:54:48.257 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- bash -c 'ceph auth get-or-create client.foo mon '"'"'allow r'"'"'' 2026-03-10T05:54:48.456 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:48.760 INFO:teuthology.orchestra.run.vm01.stdout:[client.foo] 2026-03-10T05:54:48.760 INFO:teuthology.orchestra.run.vm01.stdout: key = AQAosq9pOZnjLBAAtkNDKQSk/g1JD4uYAWzIZg== 2026-03-10T05:54:48.930 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- bash -c 'ceph orch client-keyring set client.foo label:foo --mode 770 --owner 11111:22222' 2026-03-10T05:54:49.105 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:48 vm01 ceph-mon[47299]: pgmap v50: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 2.8 KiB/s rd, 58 KiB/s wr, 7 op/s 2026-03-10T05:54:49.105 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:48 vm01 ceph-mon[47299]: mgrmap e19: vm01.wforik(active, since 72s), standbys: vm09.uxbryv 2026-03-10T05:54:49.105 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:48 vm01 ceph-mon[47299]: from='client.24343 -' entity='client.admin' cmd=[{"prefix": "orch host label add", "hostname": "vm01", "label": "foo", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:54:49.105 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:48 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:49.105 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:48 vm01 ceph-mon[47299]: Added label foo to host vm01 2026-03-10T05:54:49.105 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:48 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:54:49.105 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:48 vm01 ceph-mon[47299]: from='client.? 192.168.123.101:0/123288009' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.foo", "caps": ["mon", "allow r"]}]: dispatch 2026-03-10T05:54:49.105 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:48 vm01 ceph-mon[47299]: from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.foo", "caps": ["mon", "allow r"]}]: dispatch 2026-03-10T05:54:49.105 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:48 vm01 ceph-mon[47299]: from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.foo", "caps": ["mon", "allow r"]}]': finished 2026-03-10T05:54:49.105 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:48 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:49.105 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:48 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:49.141 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:49.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:48 vm09 ceph-mon[53707]: pgmap v50: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 2.8 KiB/s rd, 58 KiB/s wr, 7 op/s 2026-03-10T05:54:49.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:48 vm09 ceph-mon[53707]: mgrmap e19: vm01.wforik(active, since 72s), standbys: vm09.uxbryv 2026-03-10T05:54:49.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:48 vm09 ceph-mon[53707]: from='client.24343 -' entity='client.admin' cmd=[{"prefix": "orch host label add", "hostname": "vm01", "label": "foo", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:54:49.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:48 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:49.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:48 vm09 ceph-mon[53707]: Added label foo to host vm01 2026-03-10T05:54:49.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:48 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:54:49.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:48 vm09 ceph-mon[53707]: from='client.? 192.168.123.101:0/123288009' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.foo", "caps": ["mon", "allow r"]}]: dispatch 2026-03-10T05:54:49.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:48 vm09 ceph-mon[53707]: from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.foo", "caps": ["mon", "allow r"]}]: dispatch 2026-03-10T05:54:49.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:48 vm09 ceph-mon[53707]: from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.foo", "caps": ["mon", "allow r"]}]': finished 2026-03-10T05:54:49.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:48 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:49.210 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:48 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:49.565 INFO:teuthology.run_tasks:Running task exec... 2026-03-10T05:54:49.567 INFO:teuthology.task.exec:Executing custom commands... 2026-03-10T05:54:49.567 INFO:teuthology.task.exec:Running commands on role host.a host ubuntu@vm01.local 2026-03-10T05:54:49.567 DEBUG:teuthology.orchestra.run.vm01:> sudo TESTDIR=/home/ubuntu/cephtest bash -c 'while ! test -e /etc/ceph/ceph.client.foo.keyring ; do sleep 1 ; done' 2026-03-10T05:54:50.602 DEBUG:teuthology.orchestra.run.vm01:> sudo TESTDIR=/home/ubuntu/cephtest bash -c 'ls -al /etc/ceph/ceph.client.foo.keyring | grep rwxrwx---' 2026-03-10T05:54:50.632 INFO:teuthology.orchestra.run.vm01.stdout:-rwxrwx---. 1 11111 22222 83 Mar 10 05:54 /etc/ceph/ceph.client.foo.keyring 2026-03-10T05:54:50.633 DEBUG:teuthology.orchestra.run.vm01:> sudo TESTDIR=/home/ubuntu/cephtest bash -c 'ls -al /etc/ceph/ceph.client.foo.keyring | grep 11111' 2026-03-10T05:54:50.704 INFO:teuthology.orchestra.run.vm01.stdout:-rwxrwx---. 1 11111 22222 83 Mar 10 05:54 /etc/ceph/ceph.client.foo.keyring 2026-03-10T05:54:50.706 DEBUG:teuthology.orchestra.run.vm01:> sudo TESTDIR=/home/ubuntu/cephtest bash -c 'ls -al /etc/ceph/ceph.client.foo.keyring | grep 22222' 2026-03-10T05:54:50.791 INFO:teuthology.orchestra.run.vm01.stdout:-rwxrwx---. 1 11111 22222 83 Mar 10 05:54 /etc/ceph/ceph.client.foo.keyring 2026-03-10T05:54:50.793 DEBUG:teuthology.orchestra.run.vm01:> sudo TESTDIR=/home/ubuntu/cephtest bash -c 'test -e /etc/ceph/ceph.conf' 2026-03-10T05:54:50.826 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:50 vm01 ceph-mon[47299]: from='client.14576 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.foo", "placement": "label:foo", "mode": "770", "owner": "11111:22222", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:54:50.827 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:50 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:50.827 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:50 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:50.827 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:50 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:50.827 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:50 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:50.827 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:50 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:54:50.827 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:50 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "client.foo"}]: dispatch 2026-03-10T05:54:50.827 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:50 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:50.830 INFO:teuthology.run_tasks:Running task exec... 2026-03-10T05:54:50.832 INFO:teuthology.task.exec:Executing custom commands... 2026-03-10T05:54:50.832 INFO:teuthology.task.exec:Running commands on role host.b host ubuntu@vm09.local 2026-03-10T05:54:50.832 DEBUG:teuthology.orchestra.run.vm09:> sudo TESTDIR=/home/ubuntu/cephtest bash -c 'test ! -e /etc/ceph/ceph.client.foo.keyring' 2026-03-10T05:54:50.858 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:50 vm09 ceph-mon[53707]: from='client.14576 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.foo", "placement": "label:foo", "mode": "770", "owner": "11111:22222", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:54:50.858 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:50 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:50.858 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:50 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:50.858 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:50 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:50.858 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:50 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:50.858 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:50 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:54:50.858 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:50 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "client.foo"}]: dispatch 2026-03-10T05:54:50.858 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:50 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:50.861 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-10T05:54:50.864 INFO:tasks.cephadm:Running commands on role host.b host ubuntu@vm09.local 2026-03-10T05:54:50.864 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- bash -c 'ceph orch host label add `hostname` foo' 2026-03-10T05:54:51.095 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm09/config 2026-03-10T05:54:51.426 INFO:teuthology.orchestra.run.vm09.stdout:Added label foo to host vm09 2026-03-10T05:54:51.616 INFO:teuthology.run_tasks:Running task exec... 2026-03-10T05:54:51.618 INFO:teuthology.task.exec:Executing custom commands... 2026-03-10T05:54:51.618 INFO:teuthology.task.exec:Running commands on role host.b host ubuntu@vm09.local 2026-03-10T05:54:51.618 DEBUG:teuthology.orchestra.run.vm09:> sudo TESTDIR=/home/ubuntu/cephtest bash -c 'while ! test -e /etc/ceph/ceph.client.foo.keyring ; do sleep 1 ; done' 2026-03-10T05:54:51.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:51 vm09 ceph-mon[53707]: pgmap v51: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 2.3 KiB/s rd, 49 KiB/s wr, 6 op/s 2026-03-10T05:54:51.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:51 vm09 ceph-mon[53707]: Updating vm01:/etc/ceph/ceph.client.foo.keyring 2026-03-10T05:54:51.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:51 vm09 ceph-mon[53707]: Updating vm01:/var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/config/ceph.client.foo.keyring 2026-03-10T05:54:51.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:51 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:51.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:51 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:51.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:51 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:54:51.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:51 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:54:51.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:51 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T05:54:51.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:51 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:51.822 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:51 vm01 ceph-mon[47299]: pgmap v51: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 2.3 KiB/s rd, 49 KiB/s wr, 6 op/s 2026-03-10T05:54:51.822 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:51 vm01 ceph-mon[47299]: Updating vm01:/etc/ceph/ceph.client.foo.keyring 2026-03-10T05:54:51.822 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:51 vm01 ceph-mon[47299]: Updating vm01:/var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/config/ceph.client.foo.keyring 2026-03-10T05:54:51.822 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:51 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:51.822 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:51 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:51.822 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:51 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:54:51.822 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:51 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:54:51.822 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:51 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T05:54:51.822 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:51 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:52.645 DEBUG:teuthology.orchestra.run.vm09:> sudo TESTDIR=/home/ubuntu/cephtest bash -c 'ls -al /etc/ceph/ceph.client.foo.keyring | grep rwxrwx---' 2026-03-10T05:54:52.680 INFO:teuthology.orchestra.run.vm09.stdout:-rwxrwx---. 1 11111 22222 83 Mar 10 05:54 /etc/ceph/ceph.client.foo.keyring 2026-03-10T05:54:52.681 DEBUG:teuthology.orchestra.run.vm09:> sudo TESTDIR=/home/ubuntu/cephtest bash -c 'ls -al /etc/ceph/ceph.client.foo.keyring | grep 11111' 2026-03-10T05:54:52.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:52 vm01 ceph-mon[47299]: from='client.24351 -' entity='client.admin' cmd=[{"prefix": "orch host label add", "hostname": "vm09", "label": "foo", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:54:52.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:52 vm01 ceph-mon[47299]: Added label foo to host vm09 2026-03-10T05:54:52.711 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:52 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:52.711 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:52 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:52.711 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:52 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:52.711 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:52 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:52.711 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:52 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:52.711 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:52 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:54:52.711 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:52 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "client.foo"}]: dispatch 2026-03-10T05:54:52.711 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:52 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:52.711 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:52 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:52.711 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:52 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:52.711 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:52 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:54:52.711 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:52 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:54:52.760 INFO:teuthology.orchestra.run.vm09.stdout:-rwxrwx---. 1 11111 22222 83 Mar 10 05:54 /etc/ceph/ceph.client.foo.keyring 2026-03-10T05:54:52.762 DEBUG:teuthology.orchestra.run.vm09:> sudo TESTDIR=/home/ubuntu/cephtest bash -c 'ls -al /etc/ceph/ceph.client.foo.keyring | grep 22222' 2026-03-10T05:54:52.828 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:52 vm09 ceph-mon[53707]: from='client.24351 -' entity='client.admin' cmd=[{"prefix": "orch host label add", "hostname": "vm09", "label": "foo", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:54:52.828 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:52 vm09 ceph-mon[53707]: Added label foo to host vm09 2026-03-10T05:54:52.828 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:52 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:52.828 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:52 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:52.828 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:52 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:52.828 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:52 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:52.828 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:52 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:52.828 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:52 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:54:52.828 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:52 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "client.foo"}]: dispatch 2026-03-10T05:54:52.828 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:52 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:52.828 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:52 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:52.828 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:52 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:52.828 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:52 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:54:52.828 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:52 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:54:52.834 INFO:teuthology.orchestra.run.vm09.stdout:-rwxrwx---. 1 11111 22222 83 Mar 10 05:54 /etc/ceph/ceph.client.foo.keyring 2026-03-10T05:54:52.836 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-10T05:54:52.839 INFO:tasks.cephadm:Running commands on role host.b host ubuntu@vm09.local 2026-03-10T05:54:52.839 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- bash -c 'ceph orch host label rm `hostname` foo' 2026-03-10T05:54:53.084 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm09/config 2026-03-10T05:54:53.329 INFO:teuthology.orchestra.run.vm09.stdout:Removed label foo from host vm09 2026-03-10T05:54:53.486 INFO:teuthology.run_tasks:Running task exec... 2026-03-10T05:54:53.500 INFO:teuthology.task.exec:Executing custom commands... 2026-03-10T05:54:53.500 INFO:teuthology.task.exec:Running commands on role host.b host ubuntu@vm09.local 2026-03-10T05:54:53.500 DEBUG:teuthology.orchestra.run.vm09:> sudo TESTDIR=/home/ubuntu/cephtest bash -c 'while test -e /etc/ceph/ceph.client.foo.keyring ; do sleep 1 ; done' 2026-03-10T05:54:53.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:53 vm09 ceph-mon[53707]: pgmap v52: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 2.3 KiB/s rd, 49 KiB/s wr, 6 op/s 2026-03-10T05:54:53.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:53 vm09 ceph-mon[53707]: Updating vm09:/etc/ceph/ceph.client.foo.keyring 2026-03-10T05:54:53.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:53 vm09 ceph-mon[53707]: Updating vm09:/var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/config/ceph.client.foo.keyring 2026-03-10T05:54:53.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:53 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:53.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:53 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:53.710 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:53 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:53.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:53 vm01 ceph-mon[47299]: pgmap v52: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 2.3 KiB/s rd, 49 KiB/s wr, 6 op/s 2026-03-10T05:54:53.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:53 vm01 ceph-mon[47299]: Updating vm09:/etc/ceph/ceph.client.foo.keyring 2026-03-10T05:54:53.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:53 vm01 ceph-mon[47299]: Updating vm09:/var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/config/ceph.client.foo.keyring 2026-03-10T05:54:53.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:53 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:53.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:53 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:53.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:53 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:54.528 INFO:teuthology.run_tasks:Running task exec... 2026-03-10T05:54:54.540 INFO:teuthology.task.exec:Executing custom commands... 2026-03-10T05:54:54.540 INFO:teuthology.task.exec:Running commands on role host.a host ubuntu@vm01.local 2026-03-10T05:54:54.540 DEBUG:teuthology.orchestra.run.vm01:> sudo TESTDIR=/home/ubuntu/cephtest bash -c 'test -e /etc/ceph/ceph.client.foo.keyring' 2026-03-10T05:54:54.567 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-10T05:54:54.569 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm01.local 2026-03-10T05:54:54.569 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- bash -c 'ceph orch client-keyring rm client.foo' 2026-03-10T05:54:54.641 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:54 vm01 ceph-mon[47299]: from='client.24355 -' entity='client.admin' cmd=[{"prefix": "orch host label rm", "hostname": "vm09", "label": "foo", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:54:54.641 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:54 vm01 ceph-mon[47299]: Removed label foo from host vm09 2026-03-10T05:54:54.641 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:54 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:54.641 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:54 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:54.641 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:54 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:54.641 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:54 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:54:54.641 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:54 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "client.foo"}]: dispatch 2026-03-10T05:54:54.641 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:54 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:54.641 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:54 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:54.641 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:54 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:54.641 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:54 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:54:54.641 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:54 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:54:54.641 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:54 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:54.641 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:54 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:54.814 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:54.959 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:54 vm09 ceph-mon[53707]: from='client.24355 -' entity='client.admin' cmd=[{"prefix": "orch host label rm", "hostname": "vm09", "label": "foo", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:54:54.959 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:54 vm09 ceph-mon[53707]: Removed label foo from host vm09 2026-03-10T05:54:54.959 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:54 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:54.959 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:54 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:54.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:54 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:54.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:54 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:54:54.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:54 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "client.foo"}]: dispatch 2026-03-10T05:54:54.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:54 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:54.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:54 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:54.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:54 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:54.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:54 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:54:54.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:54 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:54:54.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:54 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:54.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:54 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:55.295 INFO:teuthology.run_tasks:Running task exec... 2026-03-10T05:54:55.297 INFO:teuthology.task.exec:Executing custom commands... 2026-03-10T05:54:55.297 INFO:teuthology.task.exec:Running commands on role host.a host ubuntu@vm01.local 2026-03-10T05:54:55.298 DEBUG:teuthology.orchestra.run.vm01:> sudo TESTDIR=/home/ubuntu/cephtest bash -c 'while test -e /etc/ceph/ceph.client.foo.keyring ; do sleep 1 ; done' 2026-03-10T05:54:55.366 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-10T05:54:55.369 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm01.local 2026-03-10T05:54:55.369 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- bash -c 'stat -c '"'"'%u %g'"'"' /var/log/ceph | grep '"'"'167 167'"'"'' 2026-03-10T05:54:55.575 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:55 vm01 ceph-mon[47299]: Removing vm09:/etc/ceph/ceph.client.foo.keyring 2026-03-10T05:54:55.575 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:55 vm01 ceph-mon[47299]: Removing vm09:/var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/config/ceph.client.foo.keyring 2026-03-10T05:54:55.575 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:55 vm01 ceph-mon[47299]: pgmap v53: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 2.3 KiB/s rd, 49 KiB/s wr, 6 op/s 2026-03-10T05:54:55.575 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:55 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:55.575 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:55 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:55.575 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:55 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:55.575 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:55 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:55.575 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:55 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:54:55.575 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:55 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:55.575 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:55 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:55.575 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:55 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:55.575 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:55 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:54:55.575 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:55 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:54:55.638 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:55.739 INFO:teuthology.orchestra.run.vm01.stdout:167 167 2026-03-10T05:54:55.885 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:55 vm09 ceph-mon[53707]: Removing vm09:/etc/ceph/ceph.client.foo.keyring 2026-03-10T05:54:55.885 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:55 vm09 ceph-mon[53707]: Removing vm09:/var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/config/ceph.client.foo.keyring 2026-03-10T05:54:55.885 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:55 vm09 ceph-mon[53707]: pgmap v53: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 2.3 KiB/s rd, 49 KiB/s wr, 6 op/s 2026-03-10T05:54:55.885 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:55 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:55.885 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:55 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:55.885 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:55 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:55.885 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:55 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:55.885 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:55 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:54:55.885 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:55 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:55.885 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:55 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:55.885 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:55 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:55.885 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:55 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:54:55.885 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:55 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:54:55.899 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- bash -c 'ceph orch status' 2026-03-10T05:54:56.109 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:56.400 INFO:teuthology.orchestra.run.vm01.stdout:Backend: cephadm 2026-03-10T05:54:56.400 INFO:teuthology.orchestra.run.vm01.stdout:Available: Yes 2026-03-10T05:54:56.400 INFO:teuthology.orchestra.run.vm01.stdout:Paused: No 2026-03-10T05:54:56.560 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- bash -c 'ceph orch ps' 2026-03-10T05:54:56.663 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:56 vm01 ceph-mon[47299]: from='client.14586 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring rm", "entity": "client.foo", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:54:56.663 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:56 vm01 ceph-mon[47299]: Removing vm01:/etc/ceph/ceph.client.foo.keyring 2026-03-10T05:54:56.663 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:56 vm01 ceph-mon[47299]: Removing vm01:/var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/config/ceph.client.foo.keyring 2026-03-10T05:54:56.663 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:56 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:56.663 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:56 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:56.772 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:56.959 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:56 vm09 ceph-mon[53707]: from='client.14586 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring rm", "entity": "client.foo", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:54:56.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:56 vm09 ceph-mon[53707]: Removing vm01:/etc/ceph/ceph.client.foo.keyring 2026-03-10T05:54:56.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:56 vm09 ceph-mon[53707]: Removing vm01:/var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/config/ceph.client.foo.keyring 2026-03-10T05:54:56.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:56 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:56.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:56 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:57.058 INFO:teuthology.orchestra.run.vm01.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-10T05:54:57.058 INFO:teuthology.orchestra.run.vm01.stdout:alertmanager.vm01 vm01 *:9093,9094 running (62s) 0s ago 97s 24.7M - 0.25.0 c8568f914cd2 639e5539aa67 2026-03-10T05:54:57.058 INFO:teuthology.orchestra.run.vm01.stdout:ceph-exporter.vm01 vm01 *:9926 running (105s) 0s ago 105s 9093k - 19.2.3-678-ge911bdeb 654f31e6858e 4d03abd25a35 2026-03-10T05:54:57.058 INFO:teuthology.orchestra.run.vm01.stdout:ceph-exporter.vm09 vm09 *:9926 running (77s) 0s ago 77s 6555k - 19.2.3-678-ge911bdeb 654f31e6858e 685feb193756 2026-03-10T05:54:57.058 INFO:teuthology.orchestra.run.vm01.stdout:crash.vm01 vm01 running (103s) 0s ago 103s 7654k - 19.2.3-678-ge911bdeb 654f31e6858e 1d73eef3d498 2026-03-10T05:54:57.058 INFO:teuthology.orchestra.run.vm01.stdout:crash.vm09 vm09 running (76s) 0s ago 76s 7654k - 19.2.3-678-ge911bdeb 654f31e6858e dc6057be483c 2026-03-10T05:54:57.058 INFO:teuthology.orchestra.run.vm01.stdout:grafana.vm01 vm01 *:3000 running (61s) 0s ago 90s 76.7M - 10.4.0 c8b91775d855 792467d3a04f 2026-03-10T05:54:57.058 INFO:teuthology.orchestra.run.vm01.stdout:mgr.vm01.wforik vm01 *:9283,8765,8443 running (2m) 0s ago 2m 552M - 19.2.3-678-ge911bdeb 654f31e6858e fd848140b65b 2026-03-10T05:54:57.058 INFO:teuthology.orchestra.run.vm01.stdout:mgr.vm09.uxbryv vm09 *:8443,9283,8765 running (72s) 0s ago 72s 490M - 19.2.3-678-ge911bdeb 654f31e6858e 7d2c10ecee80 2026-03-10T05:54:57.058 INFO:teuthology.orchestra.run.vm01.stdout:mon.vm01 vm01 running (2m) 0s ago 2m 53.9M 2048M 19.2.3-678-ge911bdeb 654f31e6858e d0fd780adbec 2026-03-10T05:54:57.058 INFO:teuthology.orchestra.run.vm01.stdout:mon.vm09 vm09 running (71s) 0s ago 71s 48.1M 2048M 19.2.3-678-ge911bdeb 654f31e6858e 98ffb94408d7 2026-03-10T05:54:57.058 INFO:teuthology.orchestra.run.vm01.stdout:node-exporter.vm01 vm01 *:9100 running (101s) 0s ago 101s 9559k - 1.7.0 72c9c2088986 e27f0b466530 2026-03-10T05:54:57.058 INFO:teuthology.orchestra.run.vm01.stdout:node-exporter.vm09 vm09 *:9100 running (73s) 0s ago 73s 9332k - 1.7.0 72c9c2088986 73b4f78f9107 2026-03-10T05:54:57.058 INFO:teuthology.orchestra.run.vm01.stdout:osd.0 vm09 running (38s) 0s ago 38s 58.3M 4096M 19.2.3-678-ge911bdeb 654f31e6858e 8ce2792179a6 2026-03-10T05:54:57.058 INFO:teuthology.orchestra.run.vm01.stdout:osd.1 vm01 running (38s) 0s ago 38s 58.0M 4096M 19.2.3-678-ge911bdeb 654f31e6858e b9ecca506167 2026-03-10T05:54:57.058 INFO:teuthology.orchestra.run.vm01.stdout:osd.2 vm09 running (35s) 0s ago 35s 60.6M 4096M 19.2.3-678-ge911bdeb 654f31e6858e 6cc8199a1748 2026-03-10T05:54:57.058 INFO:teuthology.orchestra.run.vm01.stdout:osd.3 vm01 running (35s) 0s ago 35s 60.6M 4096M 19.2.3-678-ge911bdeb 654f31e6858e 2c0023449970 2026-03-10T05:54:57.058 INFO:teuthology.orchestra.run.vm01.stdout:osd.4 vm09 running (32s) 0s ago 32s 33.7M 4096M 19.2.3-678-ge911bdeb 654f31e6858e c12d4f24588d 2026-03-10T05:54:57.058 INFO:teuthology.orchestra.run.vm01.stdout:osd.5 vm01 running (32s) 0s ago 32s 41.0M 4096M 19.2.3-678-ge911bdeb 654f31e6858e e85aa3b83232 2026-03-10T05:54:57.058 INFO:teuthology.orchestra.run.vm01.stdout:osd.6 vm09 running (30s) 0s ago 30s 40.9M 4096M 19.2.3-678-ge911bdeb 654f31e6858e c32ea635ccf3 2026-03-10T05:54:57.058 INFO:teuthology.orchestra.run.vm01.stdout:osd.7 vm01 running (29s) 0s ago 29s 33.1M 4096M 19.2.3-678-ge911bdeb 654f31e6858e d6d776c1ae0f 2026-03-10T05:54:57.058 INFO:teuthology.orchestra.run.vm01.stdout:prometheus.vm01 vm01 *:9095 running (59s) 0s ago 85s 32.4M - 2.51.0 1d3b7f56885b c505b49771f5 2026-03-10T05:54:57.242 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- bash -c 'ceph orch ls' 2026-03-10T05:54:57.427 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:57.698 INFO:teuthology.orchestra.run.vm01.stdout:NAME PORTS RUNNING REFRESHED AGE PLACEMENT 2026-03-10T05:54:57.698 INFO:teuthology.orchestra.run.vm01.stdout:alertmanager ?:9093,9094 1/1 0s ago 116s count:1 2026-03-10T05:54:57.698 INFO:teuthology.orchestra.run.vm01.stdout:ceph-exporter ?:9926 2/2 1s ago 118s * 2026-03-10T05:54:57.698 INFO:teuthology.orchestra.run.vm01.stdout:crash 2/2 1s ago 118s * 2026-03-10T05:54:57.698 INFO:teuthology.orchestra.run.vm01.stdout:grafana ?:3000 1/1 0s ago 117s count:1 2026-03-10T05:54:57.698 INFO:teuthology.orchestra.run.vm01.stdout:mgr 2/2 1s ago 119s count:2 2026-03-10T05:54:57.698 INFO:teuthology.orchestra.run.vm01.stdout:mon 2/2 1s ago 102s vm01:192.168.123.101=vm01;vm09:192.168.123.109=vm09;count:2 2026-03-10T05:54:57.698 INFO:teuthology.orchestra.run.vm01.stdout:node-exporter ?:9100 2/2 1s ago 117s * 2026-03-10T05:54:57.698 INFO:teuthology.orchestra.run.vm01.stdout:osd.all-available-devices 8 1s ago 62s * 2026-03-10T05:54:57.698 INFO:teuthology.orchestra.run.vm01.stdout:prometheus ?:9095 1/1 0s ago 117s count:1 2026-03-10T05:54:57.708 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:57 vm01 ceph-mon[47299]: pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 2.3 KiB/s rd, 49 KiB/s wr, 6 op/s 2026-03-10T05:54:57.708 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:57 vm01 ceph-mon[47299]: from='client.14590 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:54:57.708 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:57 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:57.708 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:57 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:57.709 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:57 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:57.709 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:57 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:54:57.709 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:57 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:57.709 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:57 vm01 ceph-mon[47299]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:54:57.853 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- bash -c 'ceph orch host ls' 2026-03-10T05:54:57.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:57 vm09 ceph-mon[53707]: pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 2.3 KiB/s rd, 49 KiB/s wr, 6 op/s 2026-03-10T05:54:57.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:57 vm09 ceph-mon[53707]: from='client.14590 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:54:57.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:57 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:57.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:57 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:57.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:57 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:54:57.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:57 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:54:57.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:57 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' 2026-03-10T05:54:57.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:57 vm09 ceph-mon[53707]: from='mgr.14217 192.168.123.101:0/2128749756' entity='mgr.vm01.wforik' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:54:58.035 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:58.284 INFO:teuthology.orchestra.run.vm01.stdout:HOST ADDR LABELS STATUS 2026-03-10T05:54:58.284 INFO:teuthology.orchestra.run.vm01.stdout:vm01 192.168.123.101 foo 2026-03-10T05:54:58.284 INFO:teuthology.orchestra.run.vm01.stdout:vm09 192.168.123.109 2026-03-10T05:54:58.284 INFO:teuthology.orchestra.run.vm01.stdout:2 hosts in cluster 2026-03-10T05:54:58.533 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- bash -c 'ceph orch device ls' 2026-03-10T05:54:58.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:58 vm01 ceph-mon[47299]: from='client.24365 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:54:58.730 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:58.959 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:58 vm09 ceph-mon[53707]: from='client.24365 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:54:59.041 INFO:teuthology.orchestra.run.vm01.stdout:HOST PATH TYPE DEVICE ID SIZE AVAILABLE REFRESHED REJECT REASONS 2026-03-10T05:54:59.041 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 27s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-10T05:54:59.041 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/vdb hdd DWNBRSTVMM01001 20.0G No 27s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T05:54:59.041 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/vdc hdd DWNBRSTVMM01002 20.0G No 27s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T05:54:59.041 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/vdd hdd DWNBRSTVMM01003 20.0G No 27s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T05:54:59.042 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/vde hdd DWNBRSTVMM01004 20.0G No 27s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T05:54:59.042 INFO:teuthology.orchestra.run.vm01.stdout:vm09 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 27s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-10T05:54:59.042 INFO:teuthology.orchestra.run.vm01.stdout:vm09 /dev/vdb hdd DWNBRSTVMM09001 20.0G No 27s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T05:54:59.042 INFO:teuthology.orchestra.run.vm01.stdout:vm09 /dev/vdc hdd DWNBRSTVMM09002 20.0G No 27s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T05:54:59.042 INFO:teuthology.orchestra.run.vm01.stdout:vm09 /dev/vdd hdd DWNBRSTVMM09003 20.0G No 27s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T05:54:59.042 INFO:teuthology.orchestra.run.vm01.stdout:vm09 /dev/vde hdd DWNBRSTVMM09004 20.0G No 27s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T05:54:59.219 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- bash -c 'ceph orch ls | grep '"'"'^osd.all-available-devices '"'"'' 2026-03-10T05:54:59.382 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:54:59.630 INFO:teuthology.orchestra.run.vm01.stdout:osd.all-available-devices 8 3s ago 64s * 2026-03-10T05:54:59.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:59 vm01 ceph-mon[47299]: from='client.24369 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:54:59.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:59 vm01 ceph-mon[47299]: pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 2.3 KiB/s rd, 49 KiB/s wr, 6 op/s 2026-03-10T05:54:59.710 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:54:59 vm01 ceph-mon[47299]: from='client.14598 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:54:59.789 DEBUG:teuthology.run_tasks:Unwinding manager cephadm 2026-03-10T05:54:59.792 INFO:tasks.cephadm:Teardown begin 2026-03-10T05:54:59.792 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T05:54:59.815 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T05:54:59.841 INFO:tasks.cephadm:Disabling cephadm mgr module 2026-03-10T05:54:59.841 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 -- ceph mgr module disable cephadm 2026-03-10T05:54:59.959 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:59 vm09 ceph-mon[53707]: from='client.24369 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:54:59.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:59 vm09 ceph-mon[53707]: pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 2.3 KiB/s rd, 49 KiB/s wr, 6 op/s 2026-03-10T05:54:59.960 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:54:59 vm09 ceph-mon[53707]: from='client.14598 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:55:00.029 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/mon.vm01/config 2026-03-10T05:55:00.050 INFO:teuthology.orchestra.run.vm01.stderr:Error: statfs /etc/ceph/ceph.client.admin.keyring: no such file or directory 2026-03-10T05:55:00.068 DEBUG:teuthology.orchestra.run:got remote process result: 125 2026-03-10T05:55:00.069 INFO:tasks.cephadm:Cleaning up testdir ceph.* files... 2026-03-10T05:55:00.069 DEBUG:teuthology.orchestra.run.vm01:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-10T05:55:00.086 DEBUG:teuthology.orchestra.run.vm09:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-10T05:55:00.100 INFO:tasks.cephadm:Stopping all daemons... 2026-03-10T05:55:00.100 INFO:tasks.cephadm.mon.vm01:Stopping mon.vm01... 2026-03-10T05:55:00.100 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4@mon.vm01 2026-03-10T05:55:00.412 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:55:00 vm01 systemd[1]: Stopping Ceph mon.vm01 for 100dc3f2-1c45-11f1-8010-b96ca19c2df4... 2026-03-10T05:55:00.412 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:55:00 vm01 ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4-mon-vm01[47273]: 2026-03-10T05:55:00.232+0000 7fc274ada640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.vm01 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T05:55:00.412 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:55:00 vm01 ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4-mon-vm01[47273]: 2026-03-10T05:55:00.232+0000 7fc274ada640 -1 mon.vm01@0(leader) e2 *** Got Signal Terminated *** 2026-03-10T05:55:00.412 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 10 05:55:00 vm01 podman[85043]: 2026-03-10 05:55:00.29922159 +0000 UTC m=+0.082629879 container died d0fd780adbeca7225f8c248f65412d9a3e15a7cd74b9bcd9cf0de6f323e725f7 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4-mon-vm01, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.license=GPLv2, CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.documentation=https://docs.ceph.com/, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, org.label-schema.build-date=20260223, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, ceph=True) 2026-03-10T05:55:00.501 DEBUG:teuthology.orchestra.run.vm01:> sudo pkill -f 'journalctl -f -n 0 -u ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4@mon.vm01.service' 2026-03-10T05:55:00.542 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T05:55:00.542 INFO:tasks.cephadm.mon.vm01:Stopped mon.vm01 2026-03-10T05:55:00.542 INFO:tasks.cephadm.mon.vm09:Stopping mon.vm09... 2026-03-10T05:55:00.542 DEBUG:teuthology.orchestra.run.vm09:> sudo systemctl stop ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4@mon.vm09 2026-03-10T05:55:00.881 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:55:00 vm09 systemd[1]: Stopping Ceph mon.vm09 for 100dc3f2-1c45-11f1-8010-b96ca19c2df4... 2026-03-10T05:55:00.881 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:55:00 vm09 ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4-mon-vm09[53683]: 2026-03-10T05:55:00.670+0000 7f4593d49640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.vm09 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T05:55:00.881 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:55:00 vm09 ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4-mon-vm09[53683]: 2026-03-10T05:55:00.670+0000 7f4593d49640 -1 mon.vm09@1(peon) e2 *** Got Signal Terminated *** 2026-03-10T05:55:00.881 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 05:55:00 vm09 podman[69736]: 2026-03-10 05:55:00.877820689 +0000 UTC m=+0.227894340 container died 98ffb94408d710f491a2c6e9425dccefa04cf62a3cb351de90bda5d4d12a35be (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4-mon-vm09, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.opencontainers.image.authors=Ceph Release Team , OSD_FLAVOR=default, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.name=CentOS Stream 9 Base Image, ceph=True, io.buildah.version=1.41.3, CEPH_REF=squid, org.label-schema.vendor=CentOS, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.build-date=20260223) 2026-03-10T05:55:01.060 DEBUG:teuthology.orchestra.run.vm09:> sudo pkill -f 'journalctl -f -n 0 -u ceph-100dc3f2-1c45-11f1-8010-b96ca19c2df4@mon.vm09.service' 2026-03-10T05:55:01.097 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T05:55:01.097 INFO:tasks.cephadm.mon.vm09:Stopped mon.vm09 2026-03-10T05:55:01.097 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 --force --keep-logs 2026-03-10T05:55:01.219 INFO:teuthology.orchestra.run.vm01.stdout:Deleting cluster with fsid: 100dc3f2-1c45-11f1-8010-b96ca19c2df4 2026-03-10T05:55:28.113 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 --force --keep-logs 2026-03-10T05:55:28.231 INFO:teuthology.orchestra.run.vm09.stdout:Deleting cluster with fsid: 100dc3f2-1c45-11f1-8010-b96ca19c2df4 2026-03-10T05:55:53.784 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T05:55:53.810 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T05:55:53.833 INFO:tasks.cephadm:Archiving crash dumps... 2026-03-10T05:55:53.833 DEBUG:teuthology.misc:Transferring archived files from vm01:/var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/crash to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/921/remote/vm01/crash 2026-03-10T05:55:53.833 DEBUG:teuthology.orchestra.run.vm01:> sudo tar c -f - -C /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/crash -- . 2026-03-10T05:55:53.873 INFO:teuthology.orchestra.run.vm01.stderr:tar: /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/crash: Cannot open: No such file or directory 2026-03-10T05:55:53.873 INFO:teuthology.orchestra.run.vm01.stderr:tar: Error is not recoverable: exiting now 2026-03-10T05:55:53.874 DEBUG:teuthology.misc:Transferring archived files from vm09:/var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/crash to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/921/remote/vm09/crash 2026-03-10T05:55:53.874 DEBUG:teuthology.orchestra.run.vm09:> sudo tar c -f - -C /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/crash -- . 2026-03-10T05:55:53.898 INFO:teuthology.orchestra.run.vm09.stderr:tar: /var/lib/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/crash: Cannot open: No such file or directory 2026-03-10T05:55:53.898 INFO:teuthology.orchestra.run.vm09.stderr:tar: Error is not recoverable: exiting now 2026-03-10T05:55:53.899 INFO:tasks.cephadm:Checking cluster log for badness... 2026-03-10T05:55:53.899 DEBUG:teuthology.orchestra.run.vm01:> sudo egrep '\[ERR\]|\[WRN\]|\[SEC\]' /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph.log | egrep CEPHADM_ | egrep -v '\(MDS_ALL_DOWN\)' | egrep -v '\(MDS_UP_LESS_THAN_MAX\)' | egrep -v CEPHADM_DAEMON_PLACE_FAIL | egrep -v CEPHADM_FAILED_DAEMON | head -n 1 2026-03-10T05:55:53.941 INFO:tasks.cephadm:Compressing logs... 2026-03-10T05:55:53.941 DEBUG:teuthology.orchestra.run.vm01:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T05:55:53.983 DEBUG:teuthology.orchestra.run.vm09:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T05:55:54.008 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-10T05:55:54.008 INFO:teuthology.orchestra.run.vm01.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-10T05:55:54.008 INFO:teuthology.orchestra.run.vm09.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-10T05:55:54.009 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-10T05:55:54.010 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-mon.vm01.log 2026-03-10T05:55:54.010 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-volume.log 2026-03-10T05:55:54.010 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph.log 2026-03-10T05:55:54.010 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/cephadm.log: gzip -5 --verbose -- /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-client.ceph-exporter.vm09.log 2026-03-10T05:55:54.011 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-volume.log: 92.0% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-10T05:55:54.011 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-mgr.vm09.uxbryv.log 2026-03-10T05:55:54.011 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-client.ceph-exporter.vm09.log: 28.2% -- replaced with /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-client.ceph-exporter.vm09.log.gz 2026-03-10T05:55:54.012 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-mon.vm09.log 2026-03-10T05:55:54.014 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-mgr.vm09.uxbryv.log: 90.9% -- replaced with /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-mgr.vm09.uxbryv.log.gz 2026-03-10T05:55:54.014 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph.audit.log 2026-03-10T05:55:54.014 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/cephadm.log: /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-mon.vm01.log: gzip -5 --verbose -- /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-mgr.vm01.wforik.log 2026-03-10T05:55:54.016 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-mon.vm09.log: 95.6% -- replaced with /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-volume.log.gz 2026-03-10T05:55:54.016 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph.log 2026-03-10T05:55:54.017 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph.log: 83.3% -- replaced with /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph.log.gz 2026-03-10T05:55:54.018 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph.audit.log: 90.8% -- replaced with /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph.audit.log.gz 2026-03-10T05:55:54.018 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph.cephadm.log 2026-03-10T05:55:54.018 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph.log: 82.0% -- replaced with /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph.log.gz 2026-03-10T05:55:54.019 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-osd.0.log 2026-03-10T05:55:54.019 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph.cephadm.log: 83.2% -- replaced with /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph.cephadm.log.gz 2026-03-10T05:55:54.019 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-osd.2.log 2026-03-10T05:55:54.020 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph.audit.log 2026-03-10T05:55:54.029 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-mgr.vm01.wforik.log: gzip -5 --verbose -- /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph.cephadm.log 2026-03-10T05:55:54.030 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-osd.0.log: gzip -5 --verbose -- /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-osd.4.log 2026-03-10T05:55:54.031 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph.audit.log: 90.7% 92.1% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-10T05:55:54.032 INFO:teuthology.orchestra.run.vm01.stderr: -- replaced with /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph.audit.log.gz 2026-03-10T05:55:54.034 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-osd.2.log: gzip -5 --verbose -- /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-osd.6.log 2026-03-10T05:55:54.036 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-osd.4.log: 92.8% -- replaced with /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-osd.0.log.gz 2026-03-10T05:55:54.038 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-volume.log 2026-03-10T05:55:54.039 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph.cephadm.log: 83.5% -- replaced with /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph.cephadm.log.gz 2026-03-10T05:55:54.042 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-client.ceph-exporter.vm01.log 2026-03-10T05:55:54.049 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-osd.6.log: 93.0% -- replaced with /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-osd.4.log.gz 2026-03-10T05:55:54.049 INFO:teuthology.orchestra.run.vm09.stderr: 92.1% -- replaced with /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-mon.vm09.log.gz 2026-03-10T05:55:54.051 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-volume.log: gzip -5 --verbose -- /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-osd.1.log 2026-03-10T05:55:54.052 INFO:teuthology.orchestra.run.vm01.stderr: 95.7%/var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-client.ceph-exporter.vm01.log: -- replaced with /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-volume.log.gz 2026-03-10T05:55:54.052 INFO:teuthology.orchestra.run.vm01.stderr: 90.9% -- replaced with /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-client.ceph-exporter.vm01.log.gz 2026-03-10T05:55:54.054 INFO:teuthology.orchestra.run.vm09.stderr: 92.8% -- replaced with /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-osd.2.log.gz 2026-03-10T05:55:54.060 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-osd.3.log 2026-03-10T05:55:54.068 INFO:teuthology.orchestra.run.vm09.stderr: 93.3% -- replaced with /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-osd.6.log.gz 2026-03-10T05:55:54.070 INFO:teuthology.orchestra.run.vm09.stderr: 2026-03-10T05:55:54.070 INFO:teuthology.orchestra.run.vm09.stderr:real 0m0.071s 2026-03-10T05:55:54.070 INFO:teuthology.orchestra.run.vm09.stderr:user 0m0.099s 2026-03-10T05:55:54.070 INFO:teuthology.orchestra.run.vm09.stderr:sys 0m0.022s 2026-03-10T05:55:54.071 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-osd.1.log: gzip -5 --verbose -- /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-osd.5.log 2026-03-10T05:55:54.078 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-osd.3.log: gzip -5 --verbose -- /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-osd.7.log 2026-03-10T05:55:54.091 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-osd.5.log: /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-osd.7.log: 93.0% -- replaced with /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-osd.1.log.gz 2026-03-10T05:55:54.097 INFO:teuthology.orchestra.run.vm01.stderr: 93.1% -- replaced with /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-osd.3.log.gz 2026-03-10T05:55:54.105 INFO:teuthology.orchestra.run.vm01.stderr: 89.9% -- replaced with /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-mgr.vm01.wforik.log.gz 2026-03-10T05:55:54.110 INFO:teuthology.orchestra.run.vm01.stderr: 93.1% -- replaced with /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-osd.5.log.gz 2026-03-10T05:55:54.112 INFO:teuthology.orchestra.run.vm01.stderr: 93.2% -- replaced with /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-osd.7.log.gz 2026-03-10T05:55:54.115 INFO:teuthology.orchestra.run.vm01.stderr: 91.6% -- replaced with /var/log/ceph/100dc3f2-1c45-11f1-8010-b96ca19c2df4/ceph-mon.vm01.log.gz 2026-03-10T05:55:54.117 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-10T05:55:54.117 INFO:teuthology.orchestra.run.vm01.stderr:real 0m0.119s 2026-03-10T05:55:54.117 INFO:teuthology.orchestra.run.vm01.stderr:user 0m0.193s 2026-03-10T05:55:54.117 INFO:teuthology.orchestra.run.vm01.stderr:sys 0m0.032s 2026-03-10T05:55:54.117 INFO:tasks.cephadm:Archiving logs... 2026-03-10T05:55:54.117 DEBUG:teuthology.misc:Transferring archived files from vm01:/var/log/ceph to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/921/remote/vm01/log 2026-03-10T05:55:54.117 DEBUG:teuthology.orchestra.run.vm01:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-10T05:55:54.194 DEBUG:teuthology.misc:Transferring archived files from vm09:/var/log/ceph to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/921/remote/vm09/log 2026-03-10T05:55:54.194 DEBUG:teuthology.orchestra.run.vm09:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-10T05:55:54.225 INFO:tasks.cephadm:Removing cluster... 2026-03-10T05:55:54.225 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 --force 2026-03-10T05:55:54.364 INFO:teuthology.orchestra.run.vm01.stdout:Deleting cluster with fsid: 100dc3f2-1c45-11f1-8010-b96ca19c2df4 2026-03-10T05:55:54.454 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 100dc3f2-1c45-11f1-8010-b96ca19c2df4 --force 2026-03-10T05:55:54.580 INFO:teuthology.orchestra.run.vm09.stdout:Deleting cluster with fsid: 100dc3f2-1c45-11f1-8010-b96ca19c2df4 2026-03-10T05:55:54.675 INFO:tasks.cephadm:Removing cephadm ... 2026-03-10T05:55:54.675 DEBUG:teuthology.orchestra.run.vm01:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-10T05:55:54.690 DEBUG:teuthology.orchestra.run.vm09:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-10T05:55:54.704 INFO:tasks.cephadm:Teardown complete 2026-03-10T05:55:54.704 DEBUG:teuthology.run_tasks:Unwinding manager clock 2026-03-10T05:55:54.707 INFO:teuthology.task.clock:Checking final clock skew... 2026-03-10T05:55:54.707 DEBUG:teuthology.orchestra.run.vm01:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T05:55:54.731 DEBUG:teuthology.orchestra.run.vm09:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T05:55:54.744 INFO:teuthology.orchestra.run.vm01.stderr:bash: line 1: ntpq: command not found 2026-03-10T05:55:54.746 INFO:teuthology.orchestra.run.vm01.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-10T05:55:54.746 INFO:teuthology.orchestra.run.vm01.stdout:=============================================================================== 2026-03-10T05:55:54.746 INFO:teuthology.orchestra.run.vm01.stdout:^- mail.trexler.at 2 6 377 57 -546us[ -528us] +/- 19ms 2026-03-10T05:55:54.746 INFO:teuthology.orchestra.run.vm01.stdout:^- a.ntp.madduck.net 2 6 377 57 -502us[ -502us] +/- 55ms 2026-03-10T05:55:54.746 INFO:teuthology.orchestra.run.vm01.stdout:^* static.222.16.42.77.clie> 2 6 377 57 +27us[ +45us] +/- 2733us 2026-03-10T05:55:54.746 INFO:teuthology.orchestra.run.vm01.stdout:^- node-3.infogral.is 2 6 377 55 -506us[ -506us] +/- 14ms 2026-03-10T05:55:54.758 INFO:teuthology.orchestra.run.vm09.stderr:bash: line 1: ntpq: command not found 2026-03-10T05:55:54.760 INFO:teuthology.orchestra.run.vm09.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-10T05:55:54.761 INFO:teuthology.orchestra.run.vm09.stdout:=============================================================================== 2026-03-10T05:55:54.761 INFO:teuthology.orchestra.run.vm09.stdout:^* static.222.16.42.77.clie> 2 6 377 58 -45us[ -214us] +/- 2741us 2026-03-10T05:55:54.761 INFO:teuthology.orchestra.run.vm09.stdout:^- node-3.infogral.is 2 6 377 57 -538us[ -538us] +/- 14ms 2026-03-10T05:55:54.761 INFO:teuthology.orchestra.run.vm09.stdout:^- mail.trexler.at 2 6 377 57 -589us[ -589us] +/- 19ms 2026-03-10T05:55:54.761 INFO:teuthology.orchestra.run.vm09.stdout:^- a.ntp.madduck.net 2 6 377 57 -371us[ -371us] +/- 55ms 2026-03-10T05:55:54.761 DEBUG:teuthology.run_tasks:Unwinding manager ansible.cephlab 2026-03-10T05:55:54.763 INFO:teuthology.task.ansible:Skipping ansible cleanup... 2026-03-10T05:55:54.763 DEBUG:teuthology.run_tasks:Unwinding manager selinux 2026-03-10T05:55:54.766 DEBUG:teuthology.run_tasks:Unwinding manager pcp 2026-03-10T05:55:54.768 DEBUG:teuthology.run_tasks:Unwinding manager internal.timer 2026-03-10T05:55:54.771 INFO:teuthology.task.internal:Duration was 422.369891 seconds 2026-03-10T05:55:54.771 DEBUG:teuthology.run_tasks:Unwinding manager internal.syslog 2026-03-10T05:55:54.773 INFO:teuthology.task.internal.syslog:Shutting down syslog monitoring... 2026-03-10T05:55:54.773 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-10T05:55:54.788 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-10T05:55:54.825 INFO:teuthology.orchestra.run.vm01.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-10T05:55:54.840 INFO:teuthology.orchestra.run.vm09.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-10T05:55:55.237 INFO:teuthology.task.internal.syslog:Checking logs for errors... 2026-03-10T05:55:55.237 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm01.local 2026-03-10T05:55:55.237 DEBUG:teuthology.orchestra.run.vm01:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-10T05:55:55.302 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm09.local 2026-03-10T05:55:55.302 DEBUG:teuthology.orchestra.run.vm09:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-10T05:55:55.329 INFO:teuthology.task.internal.syslog:Gathering journactl... 2026-03-10T05:55:55.329 DEBUG:teuthology.orchestra.run.vm01:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T05:55:55.344 DEBUG:teuthology.orchestra.run.vm09:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T05:55:55.722 INFO:teuthology.task.internal.syslog:Compressing syslogs... 2026-03-10T05:55:55.722 DEBUG:teuthology.orchestra.run.vm01:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T05:55:55.723 DEBUG:teuthology.orchestra.run.vm09:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T05:55:55.744 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T05:55:55.744 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T05:55:55.745 INFO:teuthology.orchestra.run.vm01.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-10T05:55:55.745 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T05:55:55.745 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T05:55:55.745 INFO:teuthology.orchestra.run.vm01.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-10T05:55:55.746 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T05:55:55.746 INFO:teuthology.orchestra.run.vm09.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-10T05:55:55.746 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T05:55:55.747 INFO:teuthology.orchestra.run.vm09.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-10T05:55:55.853 INFO:teuthology.orchestra.run.vm09.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 98.2% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-10T05:55:55.857 INFO:teuthology.orchestra.run.vm01.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 97.8% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-10T05:55:55.859 DEBUG:teuthology.run_tasks:Unwinding manager internal.sudo 2026-03-10T05:55:55.861 INFO:teuthology.task.internal:Restoring /etc/sudoers... 2026-03-10T05:55:55.862 DEBUG:teuthology.orchestra.run.vm01:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-10T05:55:55.924 DEBUG:teuthology.orchestra.run.vm09:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-10T05:55:55.948 DEBUG:teuthology.run_tasks:Unwinding manager internal.coredump 2026-03-10T05:55:55.951 DEBUG:teuthology.orchestra.run.vm01:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-10T05:55:55.966 DEBUG:teuthology.orchestra.run.vm09:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-10T05:55:55.987 INFO:teuthology.orchestra.run.vm01.stdout:kernel.core_pattern = core 2026-03-10T05:55:56.014 INFO:teuthology.orchestra.run.vm09.stdout:kernel.core_pattern = core 2026-03-10T05:55:56.026 DEBUG:teuthology.orchestra.run.vm01:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-10T05:55:56.056 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T05:55:56.057 DEBUG:teuthology.orchestra.run.vm09:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-10T05:55:56.080 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T05:55:56.080 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive 2026-03-10T05:55:56.083 INFO:teuthology.task.internal:Transferring archived files... 2026-03-10T05:55:56.083 DEBUG:teuthology.misc:Transferring archived files from vm01:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/921/remote/vm01 2026-03-10T05:55:56.084 DEBUG:teuthology.orchestra.run.vm01:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-10T05:55:56.123 DEBUG:teuthology.misc:Transferring archived files from vm09:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/921/remote/vm09 2026-03-10T05:55:56.123 DEBUG:teuthology.orchestra.run.vm09:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-10T05:55:56.150 INFO:teuthology.task.internal:Removing archive directory... 2026-03-10T05:55:56.150 DEBUG:teuthology.orchestra.run.vm01:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-10T05:55:56.165 DEBUG:teuthology.orchestra.run.vm09:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-10T05:55:56.205 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive_upload 2026-03-10T05:55:56.208 INFO:teuthology.task.internal:Not uploading archives. 2026-03-10T05:55:56.208 DEBUG:teuthology.run_tasks:Unwinding manager internal.base 2026-03-10T05:55:56.211 INFO:teuthology.task.internal:Tidying up after the test... 2026-03-10T05:55:56.211 DEBUG:teuthology.orchestra.run.vm01:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-10T05:55:56.220 DEBUG:teuthology.orchestra.run.vm09:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-10T05:55:56.233 INFO:teuthology.orchestra.run.vm01.stdout: 8532138 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 10 05:55 /home/ubuntu/cephtest 2026-03-10T05:55:56.261 INFO:teuthology.orchestra.run.vm09.stdout: 8532144 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 10 05:55 /home/ubuntu/cephtest 2026-03-10T05:55:56.262 DEBUG:teuthology.run_tasks:Unwinding manager console_log 2026-03-10T05:55:56.269 INFO:teuthology.run:Summary data: description: orch/cephadm/smoke-roleless/{0-distro/centos_9.stream_runc 1-start 2-services/client-keyring 3-final} duration: 422.3698914051056 owner: kyr success: true 2026-03-10T05:55:56.269 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-10T05:55:56.286 INFO:teuthology.run:pass