2026-03-06T09:56:08.759 INFO:root:teuthology version: 1.2.4.dev6+g1c580df7a 2026-03-06T09:56:08.764 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-06T09:56:08.784 INFO:teuthology.run:Config: archive_path: /archive/irq0-2026-03-06_09:41:01-orch:cephadm:osds-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/115 branch: cobaltcore-storage-v19.2.3-fasttrack-3 description: orch:cephadm:osds/{0-distro/centos_9.stream_runc 1-start 2-ops/rm-zap-flag} email: null first_in_suite: false flavor: default job_id: '115' last_in_suite: false machine_type: vps name: irq0-2026-03-06_09:41:01-orch:cephadm:osds-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps no_nested_subset: false openstack: - volumes: count: 4 size: 10 os_type: centos os_version: 9.stream overrides: admin_socket: branch: cobaltcore-storage-v19.2.3-fasttrack-3 ansible.cephlab: branch: main repo: https://github.com/kshtsk/ceph-cm-ansible.git skip_tags: nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs vars: timezone: Europe/Berlin ceph: conf: mgr: debug mgr: 20 debug ms: 1 mon: debug mon: 20 debug ms: 1 debug paxos: 20 osd: debug ms: 1 debug osd: 20 osd mclock iops capacity threshold hdd: 49000 osd shutdown pgref assert: true flavor: default log-ignorelist: - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - OSD_DOWN - CEPHADM_FAILED_DAEMON - but is still running - PG_DEGRADED log-only-match: - CEPHADM_ sha1: c24117fd5525679b799527bc1bd1f1dd0a2db5e2 ceph-deploy: conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: {} cephadm: cephadm_binary_url: https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm containers: image: harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 install: ceph: flavor: default sha1: c24117fd5525679b799527bc1bd1f1dd0a2db5e2 extra_system_packages: deb: - python3-xmltodict - s3cmd rpm: - bzip2 - perl-Test-Harness - python3-xmltodict - s3cmd repos: - name: ceph-source priority: 1 url: https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/SRPMS - name: ceph-noarch priority: 1 url: https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/noarch - name: ceph priority: 1 url: https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/x86_64 selinux: allowlist: - scontext=system_u:system_r:logrotate_t:s0 workunit: branch: tt-19.2.3-fasttrack-3-no-nvme-loop sha1: 5726a36c3452e5b72190cfceba828abc62c819b7 owner: irq0 priority: 1000 repo: https://github.com/ceph/ceph.git roles: - - host.a - client.0 - - host.b - client.1 seed: 1661 sha1: c24117fd5525679b799527bc1bd1f1dd0a2db5e2 sleep_before_teardown: 0 suite: orch:cephadm:osds suite_branch: tt-19.2.3-fasttrack-3-no-nvme-loop suite_path: /home/teuthos/src/github.com_kshtsk_ceph_5726a36c3452e5b72190cfceba828abc62c819b7/qa suite_relpath: qa suite_repo: https://github.com/kshtsk/ceph.git suite_sha1: 5726a36c3452e5b72190cfceba828abc62c819b7 targets: vm01.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBKnY8ylu7/RducHGaVykVgUBz5VQphga2X3DFjHEBaOdx8LOuRioIFKjIWLIxCej4ds0A4wxOgC19zxTPTUaoLg= vm07.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBEPoZ+5OeOpmB06CJFY5eZSDF6tFnma7/O1414hVqQ3vP3DPhsfp7WrJhSzAZEC5I/z5aI9xOE3Bqy9WsFNjU8w= tasks: - pexec: all: - sudo dnf remove nvme-cli -y - sudo dnf install runc nvmetcli nvme-cli -y - sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf - sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf - cephadm: roleless: true - cephadm.shell: host.a: - ceph orch status - ceph orch ps - ceph orch ls - ceph orch host ls - ceph orch device ls - ceph orch ls | grep '^osd.all-available-devices ' - cephadm.shell: host.a: - 'set -e set -x ceph orch ps ceph orch device ls DEVID=$(ceph device ls | grep osd.1 | awk ''{print $1}'') HOST=$(ceph orch device ls | grep "$DEVID" | awk ''{print $1}'') DEV=$(ceph orch device ls | grep "$DEVID" | awk ''{print $2}'') echo "host $HOST, dev $DEV, devid $DEVID" ceph orch osd rm --zap --replace 1 while ceph orch osd rm status | grep ^1 ; do sleep 5 ; done while ! ceph osd dump | grep osd.1 | grep "up\s*in" ; do sleep 5 ; done ' teuthology: fragments_dropped: [] meta: {} postmerge: [] teuthology_branch: clyso-debian-13 teuthology_repo: https://github.com/clyso/teuthology teuthology_sha1: 1c580df7a9c7c2aadc272da296344fd99f27c444 timestamp: 2026-03-06_09:41:01 tube: vps user: irq0 verbose: false worker_log: /home/teuthos/.teuthology/dispatcher/dispatcher.vps.43333 2026-03-06T09:56:08.784 INFO:teuthology.run:suite_path is set to /home/teuthos/src/github.com_kshtsk_ceph_5726a36c3452e5b72190cfceba828abc62c819b7/qa; will attempt to use it 2026-03-06T09:56:08.785 INFO:teuthology.run:Found tasks at /home/teuthos/src/github.com_kshtsk_ceph_5726a36c3452e5b72190cfceba828abc62c819b7/qa/tasks 2026-03-06T09:56:08.785 INFO:teuthology.run_tasks:Running task internal.save_config... 2026-03-06T09:56:08.785 INFO:teuthology.task.internal:Saving configuration 2026-03-06T09:56:08.790 INFO:teuthology.run_tasks:Running task internal.check_lock... 2026-03-06T09:56:08.791 INFO:teuthology.task.internal.check_lock:Checking locks... 2026-03-06T09:56:08.796 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm01.local', 'description': '/archive/irq0-2026-03-06_09:41:01-orch:cephadm:osds-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/115', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-06 08:55:07.132603', 'locked_by': 'irq0', 'mac_address': '52:55:00:00:00:01', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBKnY8ylu7/RducHGaVykVgUBz5VQphga2X3DFjHEBaOdx8LOuRioIFKjIWLIxCej4ds0A4wxOgC19zxTPTUaoLg='} 2026-03-06T09:56:08.801 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm07.local', 'description': '/archive/irq0-2026-03-06_09:41:01-orch:cephadm:osds-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/115', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-06 08:55:07.132223', 'locked_by': 'irq0', 'mac_address': '52:55:00:00:00:07', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBEPoZ+5OeOpmB06CJFY5eZSDF6tFnma7/O1414hVqQ3vP3DPhsfp7WrJhSzAZEC5I/z5aI9xOE3Bqy9WsFNjU8w='} 2026-03-06T09:56:08.801 INFO:teuthology.run_tasks:Running task internal.add_remotes... 2026-03-06T09:56:08.802 INFO:teuthology.task.internal:roles: ubuntu@vm01.local - ['host.a', 'client.0'] 2026-03-06T09:56:08.802 INFO:teuthology.task.internal:roles: ubuntu@vm07.local - ['host.b', 'client.1'] 2026-03-06T09:56:08.802 INFO:teuthology.run_tasks:Running task console_log... 2026-03-06T09:56:08.807 DEBUG:teuthology.task.console_log:vm01 does not support IPMI; excluding 2026-03-06T09:56:08.813 DEBUG:teuthology.task.console_log:vm07 does not support IPMI; excluding 2026-03-06T09:56:08.813 DEBUG:teuthology.exit:Installing handler: Handler(exiter=, func=.kill_console_loggers at 0x7fe01888bc70>, signals=[15]) 2026-03-06T09:56:08.813 INFO:teuthology.run_tasks:Running task internal.connect... 2026-03-06T09:56:08.814 INFO:teuthology.task.internal:Opening connections... 2026-03-06T09:56:08.814 DEBUG:teuthology.task.internal:connecting to ubuntu@vm01.local 2026-03-06T09:56:08.814 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm01.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-06T09:56:08.877 DEBUG:teuthology.task.internal:connecting to ubuntu@vm07.local 2026-03-06T09:56:08.878 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm07.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-06T09:56:08.937 INFO:teuthology.run_tasks:Running task internal.push_inventory... 2026-03-06T09:56:08.939 DEBUG:teuthology.orchestra.run.vm01:> uname -m 2026-03-06T09:56:08.975 INFO:teuthology.orchestra.run.vm01.stdout:x86_64 2026-03-06T09:56:08.975 DEBUG:teuthology.orchestra.run.vm01:> cat /etc/os-release 2026-03-06T09:56:09.030 INFO:teuthology.orchestra.run.vm01.stdout:NAME="CentOS Stream" 2026-03-06T09:56:09.030 INFO:teuthology.orchestra.run.vm01.stdout:VERSION="9" 2026-03-06T09:56:09.030 INFO:teuthology.orchestra.run.vm01.stdout:ID="centos" 2026-03-06T09:56:09.030 INFO:teuthology.orchestra.run.vm01.stdout:ID_LIKE="rhel fedora" 2026-03-06T09:56:09.030 INFO:teuthology.orchestra.run.vm01.stdout:VERSION_ID="9" 2026-03-06T09:56:09.030 INFO:teuthology.orchestra.run.vm01.stdout:PLATFORM_ID="platform:el9" 2026-03-06T09:56:09.030 INFO:teuthology.orchestra.run.vm01.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-06T09:56:09.030 INFO:teuthology.orchestra.run.vm01.stdout:ANSI_COLOR="0;31" 2026-03-06T09:56:09.030 INFO:teuthology.orchestra.run.vm01.stdout:LOGO="fedora-logo-icon" 2026-03-06T09:56:09.030 INFO:teuthology.orchestra.run.vm01.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-06T09:56:09.031 INFO:teuthology.orchestra.run.vm01.stdout:HOME_URL="https://centos.org/" 2026-03-06T09:56:09.031 INFO:teuthology.orchestra.run.vm01.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-06T09:56:09.031 INFO:teuthology.orchestra.run.vm01.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-06T09:56:09.031 INFO:teuthology.orchestra.run.vm01.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-06T09:56:09.031 INFO:teuthology.lock.ops:Updating vm01.local on lock server 2026-03-06T09:56:09.042 DEBUG:teuthology.orchestra.run.vm07:> uname -m 2026-03-06T09:56:09.060 INFO:teuthology.orchestra.run.vm07.stdout:x86_64 2026-03-06T09:56:09.060 DEBUG:teuthology.orchestra.run.vm07:> cat /etc/os-release 2026-03-06T09:56:09.118 INFO:teuthology.orchestra.run.vm07.stdout:NAME="CentOS Stream" 2026-03-06T09:56:09.118 INFO:teuthology.orchestra.run.vm07.stdout:VERSION="9" 2026-03-06T09:56:09.118 INFO:teuthology.orchestra.run.vm07.stdout:ID="centos" 2026-03-06T09:56:09.118 INFO:teuthology.orchestra.run.vm07.stdout:ID_LIKE="rhel fedora" 2026-03-06T09:56:09.118 INFO:teuthology.orchestra.run.vm07.stdout:VERSION_ID="9" 2026-03-06T09:56:09.118 INFO:teuthology.orchestra.run.vm07.stdout:PLATFORM_ID="platform:el9" 2026-03-06T09:56:09.118 INFO:teuthology.orchestra.run.vm07.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-06T09:56:09.118 INFO:teuthology.orchestra.run.vm07.stdout:ANSI_COLOR="0;31" 2026-03-06T09:56:09.118 INFO:teuthology.orchestra.run.vm07.stdout:LOGO="fedora-logo-icon" 2026-03-06T09:56:09.118 INFO:teuthology.orchestra.run.vm07.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-06T09:56:09.118 INFO:teuthology.orchestra.run.vm07.stdout:HOME_URL="https://centos.org/" 2026-03-06T09:56:09.118 INFO:teuthology.orchestra.run.vm07.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-06T09:56:09.118 INFO:teuthology.orchestra.run.vm07.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-06T09:56:09.118 INFO:teuthology.orchestra.run.vm07.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-06T09:56:09.118 INFO:teuthology.lock.ops:Updating vm07.local on lock server 2026-03-06T09:56:09.123 INFO:teuthology.run_tasks:Running task internal.serialize_remote_roles... 2026-03-06T09:56:09.124 INFO:teuthology.run_tasks:Running task internal.check_conflict... 2026-03-06T09:56:09.125 INFO:teuthology.task.internal:Checking for old test directory... 2026-03-06T09:56:09.125 DEBUG:teuthology.orchestra.run.vm01:> test '!' -e /home/ubuntu/cephtest 2026-03-06T09:56:09.127 DEBUG:teuthology.orchestra.run.vm07:> test '!' -e /home/ubuntu/cephtest 2026-03-06T09:56:09.177 INFO:teuthology.run_tasks:Running task internal.check_ceph_data... 2026-03-06T09:56:09.179 INFO:teuthology.task.internal:Checking for non-empty /var/lib/ceph... 2026-03-06T09:56:09.179 DEBUG:teuthology.orchestra.run.vm01:> test -z $(ls -A /var/lib/ceph) 2026-03-06T09:56:09.182 DEBUG:teuthology.orchestra.run.vm07:> test -z $(ls -A /var/lib/ceph) 2026-03-06T09:56:09.194 INFO:teuthology.orchestra.run.vm01.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-06T09:56:09.236 INFO:teuthology.orchestra.run.vm07.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-06T09:56:09.236 INFO:teuthology.run_tasks:Running task internal.vm_setup... 2026-03-06T09:56:09.245 DEBUG:teuthology.orchestra.run.vm01:> test -e /ceph-qa-ready 2026-03-06T09:56:09.259 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T09:56:09.452 DEBUG:teuthology.orchestra.run.vm07:> test -e /ceph-qa-ready 2026-03-06T09:56:09.470 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T09:56:09.651 INFO:teuthology.run_tasks:Running task internal.base... 2026-03-06T09:56:09.652 INFO:teuthology.task.internal:Creating test directory... 2026-03-06T09:56:09.652 DEBUG:teuthology.orchestra.run.vm01:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-06T09:56:09.654 DEBUG:teuthology.orchestra.run.vm07:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-06T09:56:09.668 INFO:teuthology.run_tasks:Running task internal.archive_upload... 2026-03-06T09:56:09.669 INFO:teuthology.run_tasks:Running task internal.archive... 2026-03-06T09:56:09.670 INFO:teuthology.task.internal:Creating archive directory... 2026-03-06T09:56:09.670 DEBUG:teuthology.orchestra.run.vm01:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-06T09:56:09.708 DEBUG:teuthology.orchestra.run.vm07:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-06T09:56:09.727 INFO:teuthology.run_tasks:Running task internal.coredump... 2026-03-06T09:56:09.728 INFO:teuthology.task.internal:Enabling coredump saving... 2026-03-06T09:56:09.729 DEBUG:teuthology.orchestra.run.vm01:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-06T09:56:09.777 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T09:56:09.777 DEBUG:teuthology.orchestra.run.vm07:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-06T09:56:09.792 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T09:56:09.792 DEBUG:teuthology.orchestra.run.vm01:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-06T09:56:09.819 DEBUG:teuthology.orchestra.run.vm07:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-06T09:56:09.841 INFO:teuthology.orchestra.run.vm01.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-06T09:56:09.852 INFO:teuthology.orchestra.run.vm01.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-06T09:56:09.858 INFO:teuthology.orchestra.run.vm07.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-06T09:56:09.868 INFO:teuthology.orchestra.run.vm07.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-06T09:56:09.870 INFO:teuthology.run_tasks:Running task internal.sudo... 2026-03-06T09:56:09.878 INFO:teuthology.task.internal:Configuring sudo... 2026-03-06T09:56:09.878 DEBUG:teuthology.orchestra.run.vm01:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-06T09:56:09.895 DEBUG:teuthology.orchestra.run.vm07:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-06T09:56:09.938 INFO:teuthology.run_tasks:Running task internal.syslog... 2026-03-06T09:56:09.940 INFO:teuthology.task.internal.syslog:Starting syslog monitoring... 2026-03-06T09:56:09.940 DEBUG:teuthology.orchestra.run.vm01:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-06T09:56:09.959 DEBUG:teuthology.orchestra.run.vm07:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-06T09:56:09.998 DEBUG:teuthology.orchestra.run.vm01:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-06T09:56:10.035 DEBUG:teuthology.orchestra.run.vm01:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-06T09:56:10.094 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T09:56:10.094 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-06T09:56:10.157 DEBUG:teuthology.orchestra.run.vm07:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-06T09:56:10.185 DEBUG:teuthology.orchestra.run.vm07:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-06T09:56:10.247 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-06T09:56:10.247 DEBUG:teuthology.orchestra.run.vm07:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-06T09:56:10.311 DEBUG:teuthology.orchestra.run.vm01:> sudo service rsyslog restart 2026-03-06T09:56:10.314 DEBUG:teuthology.orchestra.run.vm07:> sudo service rsyslog restart 2026-03-06T09:56:10.343 INFO:teuthology.orchestra.run.vm01.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-06T09:56:10.382 INFO:teuthology.orchestra.run.vm07.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-06T09:56:10.722 INFO:teuthology.run_tasks:Running task internal.timer... 2026-03-06T09:56:10.724 INFO:teuthology.task.internal:Starting timer... 2026-03-06T09:56:10.724 INFO:teuthology.run_tasks:Running task pcp... 2026-03-06T09:56:10.726 INFO:teuthology.run_tasks:Running task selinux... 2026-03-06T09:56:10.728 DEBUG:teuthology.task:Applying overrides for task selinux: {'allowlist': ['scontext=system_u:system_r:logrotate_t:s0']} 2026-03-06T09:56:10.728 INFO:teuthology.task.selinux:Excluding vm01: VMs are not yet supported 2026-03-06T09:56:10.729 INFO:teuthology.task.selinux:Excluding vm07: VMs are not yet supported 2026-03-06T09:56:10.729 DEBUG:teuthology.task.selinux:Getting current SELinux state 2026-03-06T09:56:10.729 DEBUG:teuthology.task.selinux:Existing SELinux modes: {} 2026-03-06T09:56:10.729 INFO:teuthology.task.selinux:Putting SELinux into permissive mode 2026-03-06T09:56:10.729 INFO:teuthology.run_tasks:Running task ansible.cephlab... 2026-03-06T09:56:10.730 DEBUG:teuthology.task:Applying overrides for task ansible.cephlab: {'branch': 'main', 'repo': 'https://github.com/kshtsk/ceph-cm-ansible.git', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'timezone': 'Europe/Berlin'}} 2026-03-06T09:56:10.730 DEBUG:teuthology.repo_utils:Setting repo remote to https://github.com/kshtsk/ceph-cm-ansible.git 2026-03-06T09:56:10.732 INFO:teuthology.repo_utils:Fetching github.com_kshtsk_ceph-cm-ansible_main from origin 2026-03-06T09:56:11.343 DEBUG:teuthology.repo_utils:Resetting repo at /home/teuthos/src/github.com_kshtsk_ceph-cm-ansible_main to origin/main 2026-03-06T09:56:11.348 INFO:teuthology.task.ansible:Playbook: [{'import_playbook': 'ansible_managed.yml'}, {'import_playbook': 'teuthology.yml'}, {'hosts': 'testnodes', 'tasks': [{'set_fact': {'ran_from_cephlab_playbook': True}}]}, {'import_playbook': 'testnodes.yml'}, {'import_playbook': 'container-host.yml'}, {'import_playbook': 'cobbler.yml'}, {'import_playbook': 'paddles.yml'}, {'import_playbook': 'pulpito.yml'}, {'hosts': 'testnodes', 'become': True, 'tasks': [{'name': 'Touch /ceph-qa-ready', 'file': {'path': '/ceph-qa-ready', 'state': 'touch'}, 'when': 'ran_from_cephlab_playbook|bool'}]}] 2026-03-06T09:56:11.349 DEBUG:teuthology.task.ansible:Running ansible-playbook -v --extra-vars '{"ansible_ssh_user": "ubuntu", "timezone": "Europe/Berlin"}' -i /tmp/teuth_ansible_inventorypqvegue0 --limit vm01.local,vm07.local /home/teuthos/src/github.com_kshtsk_ceph-cm-ansible_main/cephlab.yml --skip-tags nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs 2026-03-06T09:58:20.948 DEBUG:teuthology.task.ansible:Reconnecting to [Remote(name='ubuntu@vm01.local'), Remote(name='ubuntu@vm07.local')] 2026-03-06T09:58:20.948 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm01.local' 2026-03-06T09:58:20.949 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm01.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-06T09:58:21.012 DEBUG:teuthology.orchestra.run.vm01:> true 2026-03-06T09:58:21.091 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm01.local' 2026-03-06T09:58:21.091 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm07.local' 2026-03-06T09:58:21.091 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm07.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-06T09:58:21.154 DEBUG:teuthology.orchestra.run.vm07:> true 2026-03-06T09:58:21.238 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm07.local' 2026-03-06T09:58:21.238 INFO:teuthology.run_tasks:Running task clock... 2026-03-06T09:58:21.241 INFO:teuthology.task.clock:Syncing clocks and checking initial clock skew... 2026-03-06T09:58:21.241 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-06T09:58:21.241 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-06T09:58:21.244 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-06T09:58:21.244 DEBUG:teuthology.orchestra.run.vm07:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-06T09:58:21.285 INFO:teuthology.orchestra.run.vm01.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-06T09:58:21.302 INFO:teuthology.orchestra.run.vm01.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-06T09:58:21.309 INFO:teuthology.orchestra.run.vm07.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-06T09:58:21.322 INFO:teuthology.orchestra.run.vm07.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-06T09:58:21.335 INFO:teuthology.orchestra.run.vm01.stderr:sudo: ntpd: command not found 2026-03-06T09:58:21.346 INFO:teuthology.orchestra.run.vm07.stderr:sudo: ntpd: command not found 2026-03-06T09:58:21.347 INFO:teuthology.orchestra.run.vm01.stdout:506 Cannot talk to daemon 2026-03-06T09:58:21.361 INFO:teuthology.orchestra.run.vm07.stdout:506 Cannot talk to daemon 2026-03-06T09:58:21.366 INFO:teuthology.orchestra.run.vm01.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-06T09:58:21.377 INFO:teuthology.orchestra.run.vm07.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-06T09:58:21.385 INFO:teuthology.orchestra.run.vm01.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-06T09:58:21.391 INFO:teuthology.orchestra.run.vm07.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-06T09:58:21.436 INFO:teuthology.orchestra.run.vm01.stderr:bash: line 1: ntpq: command not found 2026-03-06T09:58:21.436 INFO:teuthology.orchestra.run.vm07.stderr:bash: line 1: ntpq: command not found 2026-03-06T09:58:22.174 INFO:teuthology.orchestra.run.vm01.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-06T09:58:22.174 INFO:teuthology.orchestra.run.vm01.stdout:=============================================================================== 2026-03-06T09:58:22.174 INFO:teuthology.orchestra.run.vm01.stdout:^? 158.101.188.125 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-06T09:58:22.174 INFO:teuthology.orchestra.run.vm01.stdout:^? red-pelican-63749.zap.cl> 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-06T09:58:22.174 INFO:teuthology.orchestra.run.vm01.stdout:^? x1.ncomputers.org 2 6 1 1 +895us[ +895us] +/- 46ms 2026-03-06T09:58:22.174 INFO:teuthology.orchestra.run.vm01.stdout:^? listserver.trexler.at 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-06T09:58:22.175 INFO:teuthology.orchestra.run.vm07.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-06T09:58:22.175 INFO:teuthology.orchestra.run.vm07.stdout:=============================================================================== 2026-03-06T09:58:22.175 INFO:teuthology.orchestra.run.vm07.stdout:^? listserver.trexler.at 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-06T09:58:22.175 INFO:teuthology.orchestra.run.vm07.stdout:^? 158.101.188.125 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-06T09:58:22.175 INFO:teuthology.orchestra.run.vm07.stdout:^? red-pelican-63749.zap.cl> 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-06T09:58:22.175 INFO:teuthology.orchestra.run.vm07.stdout:^? x1.ncomputers.org 2 6 1 1 -549us[ -549us] +/- 44ms 2026-03-06T09:58:22.175 INFO:teuthology.run_tasks:Running task pexec... 2026-03-06T09:58:22.179 INFO:teuthology.task.pexec:Executing custom commands... 2026-03-06T09:58:22.179 DEBUG:teuthology.orchestra.run.vm01:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-06T09:58:22.179 DEBUG:teuthology.orchestra.run.vm07:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-06T09:58:22.217 DEBUG:teuthology.task.pexec:ubuntu@vm07.local< sudo dnf remove nvme-cli -y 2026-03-06T09:58:22.217 DEBUG:teuthology.task.pexec:ubuntu@vm07.local< sudo dnf install runc nvmetcli nvme-cli -y 2026-03-06T09:58:22.217 DEBUG:teuthology.task.pexec:ubuntu@vm07.local< sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-06T09:58:22.217 DEBUG:teuthology.task.pexec:ubuntu@vm07.local< sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-06T09:58:22.217 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm07.local 2026-03-06T09:58:22.217 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-06T09:58:22.217 INFO:teuthology.task.pexec:sudo dnf install runc nvmetcli nvme-cli -y 2026-03-06T09:58:22.217 INFO:teuthology.task.pexec:sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-06T09:58:22.217 INFO:teuthology.task.pexec:sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-06T09:58:22.218 DEBUG:teuthology.task.pexec:ubuntu@vm01.local< sudo dnf remove nvme-cli -y 2026-03-06T09:58:22.218 DEBUG:teuthology.task.pexec:ubuntu@vm01.local< sudo dnf install runc nvmetcli nvme-cli -y 2026-03-06T09:58:22.218 DEBUG:teuthology.task.pexec:ubuntu@vm01.local< sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-06T09:58:22.218 DEBUG:teuthology.task.pexec:ubuntu@vm01.local< sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-06T09:58:22.218 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm01.local 2026-03-06T09:58:22.218 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-06T09:58:22.218 INFO:teuthology.task.pexec:sudo dnf install runc nvmetcli nvme-cli -y 2026-03-06T09:58:22.218 INFO:teuthology.task.pexec:sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-06T09:58:22.218 INFO:teuthology.task.pexec:sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-06T09:58:22.418 INFO:teuthology.orchestra.run.vm07.stdout:No match for argument: nvme-cli 2026-03-06T09:58:22.418 INFO:teuthology.orchestra.run.vm07.stderr:No packages marked for removal. 2026-03-06T09:58:22.421 INFO:teuthology.orchestra.run.vm07.stdout:Dependencies resolved. 2026-03-06T09:58:22.421 INFO:teuthology.orchestra.run.vm07.stdout:Nothing to do. 2026-03-06T09:58:22.421 INFO:teuthology.orchestra.run.vm07.stdout:Complete! 2026-03-06T09:58:22.445 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: nvme-cli 2026-03-06T09:58:22.446 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T09:58:22.448 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T09:58:22.449 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T09:58:22.449 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T09:58:22.795 INFO:teuthology.orchestra.run.vm07.stdout:Last metadata expiration check: 0:01:25 ago on Fri 06 Mar 2026 09:56:57 AM CET. 2026-03-06T09:58:22.897 INFO:teuthology.orchestra.run.vm07.stdout:Dependencies resolved. 2026-03-06T09:58:22.898 INFO:teuthology.orchestra.run.vm07.stdout:================================================================================ 2026-03-06T09:58:22.898 INFO:teuthology.orchestra.run.vm07.stdout: Package Arch Version Repository Size 2026-03-06T09:58:22.898 INFO:teuthology.orchestra.run.vm07.stdout:================================================================================ 2026-03-06T09:58:22.898 INFO:teuthology.orchestra.run.vm07.stdout:Installing: 2026-03-06T09:58:22.898 INFO:teuthology.orchestra.run.vm07.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-06T09:58:22.898 INFO:teuthology.orchestra.run.vm07.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-06T09:58:22.898 INFO:teuthology.orchestra.run.vm07.stdout: runc x86_64 4:1.4.0-2.el9 appstream 4.0 M 2026-03-06T09:58:22.898 INFO:teuthology.orchestra.run.vm07.stdout:Installing dependencies: 2026-03-06T09:58:22.898 INFO:teuthology.orchestra.run.vm07.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-06T09:58:22.898 INFO:teuthology.orchestra.run.vm07.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-06T09:58:22.898 INFO:teuthology.orchestra.run.vm07.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-06T09:58:22.898 INFO:teuthology.orchestra.run.vm07.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-06T09:58:22.898 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-06T09:58:22.898 INFO:teuthology.orchestra.run.vm07.stdout:Transaction Summary 2026-03-06T09:58:22.898 INFO:teuthology.orchestra.run.vm07.stdout:================================================================================ 2026-03-06T09:58:22.898 INFO:teuthology.orchestra.run.vm07.stdout:Install 7 Packages 2026-03-06T09:58:22.898 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-06T09:58:22.898 INFO:teuthology.orchestra.run.vm07.stdout:Total download size: 6.3 M 2026-03-06T09:58:22.898 INFO:teuthology.orchestra.run.vm07.stdout:Installed size: 24 M 2026-03-06T09:58:22.898 INFO:teuthology.orchestra.run.vm07.stdout:Downloading Packages: 2026-03-06T09:58:22.925 INFO:teuthology.orchestra.run.vm01.stdout:Last metadata expiration check: 0:01:24 ago on Fri 06 Mar 2026 09:56:58 AM CET. 2026-03-06T09:58:23.039 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T09:58:23.040 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T09:58:23.040 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-06T09:58:23.040 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T09:58:23.040 INFO:teuthology.orchestra.run.vm01.stdout:Installing: 2026-03-06T09:58:23.040 INFO:teuthology.orchestra.run.vm01.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-06T09:58:23.040 INFO:teuthology.orchestra.run.vm01.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-06T09:58:23.040 INFO:teuthology.orchestra.run.vm01.stdout: runc x86_64 4:1.4.0-2.el9 appstream 4.0 M 2026-03-06T09:58:23.040 INFO:teuthology.orchestra.run.vm01.stdout:Installing dependencies: 2026-03-06T09:58:23.040 INFO:teuthology.orchestra.run.vm01.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-06T09:58:23.040 INFO:teuthology.orchestra.run.vm01.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-06T09:58:23.040 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-06T09:58:23.040 INFO:teuthology.orchestra.run.vm01.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-06T09:58:23.040 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T09:58:23.040 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-06T09:58:23.040 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T09:58:23.040 INFO:teuthology.orchestra.run.vm01.stdout:Install 7 Packages 2026-03-06T09:58:23.040 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T09:58:23.040 INFO:teuthology.orchestra.run.vm01.stdout:Total download size: 6.3 M 2026-03-06T09:58:23.040 INFO:teuthology.orchestra.run.vm01.stdout:Installed size: 24 M 2026-03-06T09:58:23.040 INFO:teuthology.orchestra.run.vm01.stdout:Downloading Packages: 2026-03-06T09:58:23.736 INFO:teuthology.orchestra.run.vm01.stdout:(1/7): nvmetcli-0.8-3.el9.noarch.rpm 317 kB/s | 44 kB 00:00 2026-03-06T09:58:23.816 INFO:teuthology.orchestra.run.vm01.stdout:(2/7): python3-kmod-0.9-32.el9.x86_64.rpm 1.0 MB/s | 84 kB 00:00 2026-03-06T09:58:23.877 INFO:teuthology.orchestra.run.vm01.stdout:(3/7): nvme-cli-2.16-1.el9.x86_64.rpm 4.1 MB/s | 1.2 MB 00:00 2026-03-06T09:58:23.901 INFO:teuthology.orchestra.run.vm01.stdout:(4/7): python3-pyparsing-2.4.7-9.el9.noarch.rpm 1.7 MB/s | 150 kB 00:00 2026-03-06T09:58:24.030 INFO:teuthology.orchestra.run.vm01.stdout:(5/7): python3-urwid-2.1.2-4.el9.x86_64.rpm 5.3 MB/s | 837 kB 00:00 2026-03-06T09:58:24.074 INFO:teuthology.orchestra.run.vm01.stdout:(6/7): python3-configshell-1.1.30-1.el9.noarch. 151 kB/s | 72 kB 00:00 2026-03-06T09:58:24.477 INFO:teuthology.orchestra.run.vm01.stdout:(7/7): runc-1.4.0-2.el9.x86_64.rpm 6.9 MB/s | 4.0 MB 00:00 2026-03-06T09:58:24.477 INFO:teuthology.orchestra.run.vm01.stdout:-------------------------------------------------------------------------------- 2026-03-06T09:58:24.477 INFO:teuthology.orchestra.run.vm01.stdout:Total 4.4 MB/s | 6.3 MB 00:01 2026-03-06T09:58:24.584 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-06T09:58:24.596 INFO:teuthology.orchestra.run.vm07.stdout:(1/7): nvmetcli-0.8-3.el9.noarch.rpm 62 kB/s | 44 kB 00:00 2026-03-06T09:58:24.596 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-06T09:58:24.596 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-06T09:58:24.621 INFO:teuthology.orchestra.run.vm07.stdout:(2/7): python3-configshell-1.1.30-1.el9.noarch. 98 kB/s | 72 kB 00:00 2026-03-06T09:58:24.683 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-06T09:58:24.683 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-06T09:58:24.774 INFO:teuthology.orchestra.run.vm07.stdout:(3/7): python3-kmod-0.9-32.el9.x86_64.rpm 472 kB/s | 84 kB 00:00 2026-03-06T09:58:24.808 INFO:teuthology.orchestra.run.vm07.stdout:(4/7): python3-pyparsing-2.4.7-9.el9.noarch.rpm 803 kB/s | 150 kB 00:00 2026-03-06T09:58:24.890 INFO:teuthology.orchestra.run.vm07.stdout:(5/7): nvme-cli-2.16-1.el9.x86_64.rpm 1.2 MB/s | 1.2 MB 00:00 2026-03-06T09:58:24.906 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-06T09:58:24.918 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/7 2026-03-06T09:58:24.929 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/7 2026-03-06T09:58:24.937 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/7 2026-03-06T09:58:24.945 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/7 2026-03-06T09:58:24.947 INFO:teuthology.orchestra.run.vm01.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/7 2026-03-06T09:58:25.001 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/7 2026-03-06T09:58:25.100 INFO:teuthology.orchestra.run.vm07.stdout:(6/7): python3-urwid-2.1.2-4.el9.x86_64.rpm 2.5 MB/s | 837 kB 00:00 2026-03-06T09:58:25.177 INFO:teuthology.orchestra.run.vm01.stdout: Installing : runc-4:1.4.0-2.el9.x86_64 6/7 2026-03-06T09:58:25.181 INFO:teuthology.orchestra.run.vm01.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 7/7 2026-03-06T09:58:25.208 INFO:teuthology.orchestra.run.vm07.stdout:(7/7): runc-1.4.0-2.el9.x86_64.rpm 9.9 MB/s | 4.0 MB 00:00 2026-03-06T09:58:25.208 INFO:teuthology.orchestra.run.vm07.stdout:-------------------------------------------------------------------------------- 2026-03-06T09:58:25.208 INFO:teuthology.orchestra.run.vm07.stdout:Total 2.7 MB/s | 6.3 MB 00:02 2026-03-06T09:58:25.293 INFO:teuthology.orchestra.run.vm07.stdout:Running transaction check 2026-03-06T09:58:25.303 INFO:teuthology.orchestra.run.vm07.stdout:Transaction check succeeded. 2026-03-06T09:58:25.303 INFO:teuthology.orchestra.run.vm07.stdout:Running transaction test 2026-03-06T09:58:25.368 INFO:teuthology.orchestra.run.vm07.stdout:Transaction test succeeded. 2026-03-06T09:58:25.368 INFO:teuthology.orchestra.run.vm07.stdout:Running transaction 2026-03-06T09:58:25.538 INFO:teuthology.orchestra.run.vm07.stdout: Preparing : 1/1 2026-03-06T09:58:25.548 INFO:teuthology.orchestra.run.vm07.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/7 2026-03-06T09:58:25.559 INFO:teuthology.orchestra.run.vm07.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/7 2026-03-06T09:58:25.568 INFO:teuthology.orchestra.run.vm07.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/7 2026-03-06T09:58:25.579 INFO:teuthology.orchestra.run.vm07.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/7 2026-03-06T09:58:25.582 INFO:teuthology.orchestra.run.vm07.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/7 2026-03-06T09:58:25.633 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 7/7 2026-03-06T09:58:25.633 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-06T09:58:25.633 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T09:58:25.635 INFO:teuthology.orchestra.run.vm07.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/7 2026-03-06T09:58:25.774 INFO:teuthology.orchestra.run.vm07.stdout: Installing : runc-4:1.4.0-2.el9.x86_64 6/7 2026-03-06T09:58:25.779 INFO:teuthology.orchestra.run.vm07.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 7/7 2026-03-06T09:58:26.164 INFO:teuthology.orchestra.run.vm07.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 7/7 2026-03-06T09:58:26.165 INFO:teuthology.orchestra.run.vm07.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-06T09:58:26.165 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-06T09:58:26.230 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/7 2026-03-06T09:58:26.230 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/7 2026-03-06T09:58:26.230 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/7 2026-03-06T09:58:26.230 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/7 2026-03-06T09:58:26.230 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/7 2026-03-06T09:58:26.230 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/7 2026-03-06T09:58:26.321 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : runc-4:1.4.0-2.el9.x86_64 7/7 2026-03-06T09:58:26.321 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T09:58:26.321 INFO:teuthology.orchestra.run.vm01.stdout:Installed: 2026-03-06T09:58:26.321 INFO:teuthology.orchestra.run.vm01.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-06T09:58:26.321 INFO:teuthology.orchestra.run.vm01.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-06T09:58:26.321 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-06T09:58:26.321 INFO:teuthology.orchestra.run.vm01.stdout: runc-4:1.4.0-2.el9.x86_64 2026-03-06T09:58:26.321 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T09:58:26.321 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T09:58:26.423 DEBUG:teuthology.parallel:result is None 2026-03-06T09:58:26.750 INFO:teuthology.orchestra.run.vm07.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/7 2026-03-06T09:58:26.750 INFO:teuthology.orchestra.run.vm07.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/7 2026-03-06T09:58:26.750 INFO:teuthology.orchestra.run.vm07.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/7 2026-03-06T09:58:26.750 INFO:teuthology.orchestra.run.vm07.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/7 2026-03-06T09:58:26.750 INFO:teuthology.orchestra.run.vm07.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/7 2026-03-06T09:58:26.750 INFO:teuthology.orchestra.run.vm07.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/7 2026-03-06T09:58:26.837 INFO:teuthology.orchestra.run.vm07.stdout: Verifying : runc-4:1.4.0-2.el9.x86_64 7/7 2026-03-06T09:58:26.838 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-06T09:58:26.838 INFO:teuthology.orchestra.run.vm07.stdout:Installed: 2026-03-06T09:58:26.838 INFO:teuthology.orchestra.run.vm07.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-06T09:58:26.838 INFO:teuthology.orchestra.run.vm07.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-06T09:58:26.838 INFO:teuthology.orchestra.run.vm07.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-06T09:58:26.838 INFO:teuthology.orchestra.run.vm07.stdout: runc-4:1.4.0-2.el9.x86_64 2026-03-06T09:58:26.838 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-06T09:58:26.838 INFO:teuthology.orchestra.run.vm07.stdout:Complete! 2026-03-06T09:58:26.947 DEBUG:teuthology.parallel:result is None 2026-03-06T09:58:26.947 INFO:teuthology.run_tasks:Running task cephadm... 2026-03-06T09:58:26.991 INFO:tasks.cephadm:Config: {'roleless': True, 'conf': {'mgr': {'debug mgr': 20, 'debug ms': 1}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000, 'osd shutdown pgref assert': True}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', 'OSD_DOWN', 'CEPHADM_FAILED_DAEMON', 'but is still running', 'PG_DEGRADED'], 'log-only-match': ['CEPHADM_'], 'sha1': 'c24117fd5525679b799527bc1bd1f1dd0a2db5e2', 'cephadm_binary_url': 'https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm', 'containers': {'image': 'harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3'}} 2026-03-06T09:58:26.991 INFO:tasks.cephadm:Provided image contains tag or digest, using it as is 2026-03-06T09:58:26.991 INFO:tasks.cephadm:Cluster image is harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 2026-03-06T09:58:26.991 INFO:tasks.cephadm:Cluster fsid is a3ab6672-193a-11f1-b81f-a119763c7190 2026-03-06T09:58:26.991 INFO:tasks.cephadm:Choosing monitor IPs and ports... 2026-03-06T09:58:26.991 INFO:tasks.cephadm:No mon roles; fabricating mons 2026-03-06T09:58:26.991 INFO:tasks.cephadm:Monitor IPs: {'mon.vm01': '192.168.123.101', 'mon.vm07': '192.168.123.107'} 2026-03-06T09:58:26.991 INFO:tasks.cephadm:Normalizing hostnames... 2026-03-06T09:58:26.991 DEBUG:teuthology.orchestra.run.vm01:> sudo hostname $(hostname -s) 2026-03-06T09:58:27.033 DEBUG:teuthology.orchestra.run.vm07:> sudo hostname $(hostname -s) 2026-03-06T09:58:27.069 INFO:tasks.cephadm:Downloading cephadm from url: https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm 2026-03-06T09:58:27.069 DEBUG:teuthology.orchestra.run.vm01:> curl --silent -L https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-06T09:58:28.172 INFO:teuthology.orchestra.run.vm01.stdout:-rw-r--r--. 1 ubuntu ubuntu 787672 Mar 6 09:58 /home/ubuntu/cephtest/cephadm 2026-03-06T09:58:28.172 DEBUG:teuthology.orchestra.run.vm07:> curl --silent -L https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-06T09:58:29.237 INFO:teuthology.orchestra.run.vm07.stdout:-rw-r--r--. 1 ubuntu ubuntu 787672 Mar 6 09:58 /home/ubuntu/cephtest/cephadm 2026-03-06T09:58:29.237 DEBUG:teuthology.orchestra.run.vm01:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-06T09:58:29.264 DEBUG:teuthology.orchestra.run.vm07:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-06T09:58:29.290 INFO:tasks.cephadm:Pulling image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 on all hosts... 2026-03-06T09:58:29.290 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 pull 2026-03-06T09:58:29.306 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 pull 2026-03-06T09:58:29.696 INFO:teuthology.orchestra.run.vm07.stderr:Pulling container image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3... 2026-03-06T09:58:29.742 INFO:teuthology.orchestra.run.vm01.stderr:Pulling container image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3... 2026-03-06T09:58:49.988 INFO:teuthology.orchestra.run.vm07.stdout:{ 2026-03-06T09:58:49.988 INFO:teuthology.orchestra.run.vm07.stdout: "ceph_version": "ceph version 19.2.3-47-gc24117fd552 (c24117fd5525679b799527bc1bd1f1dd0a2db5e2) squid (stable)", 2026-03-06T09:58:49.988 INFO:teuthology.orchestra.run.vm07.stdout: "image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", 2026-03-06T09:58:49.988 INFO:teuthology.orchestra.run.vm07.stdout: "repo_digests": [ 2026-03-06T09:58:49.988 INFO:teuthology.orchestra.run.vm07.stdout: "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b" 2026-03-06T09:58:49.988 INFO:teuthology.orchestra.run.vm07.stdout: ] 2026-03-06T09:58:49.988 INFO:teuthology.orchestra.run.vm07.stdout:} 2026-03-06T09:59:08.025 INFO:teuthology.orchestra.run.vm01.stdout:{ 2026-03-06T09:59:08.025 INFO:teuthology.orchestra.run.vm01.stdout: "ceph_version": "ceph version 19.2.3-47-gc24117fd552 (c24117fd5525679b799527bc1bd1f1dd0a2db5e2) squid (stable)", 2026-03-06T09:59:08.025 INFO:teuthology.orchestra.run.vm01.stdout: "image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", 2026-03-06T09:59:08.025 INFO:teuthology.orchestra.run.vm01.stdout: "repo_digests": [ 2026-03-06T09:59:08.025 INFO:teuthology.orchestra.run.vm01.stdout: "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b" 2026-03-06T09:59:08.025 INFO:teuthology.orchestra.run.vm01.stdout: ] 2026-03-06T09:59:08.025 INFO:teuthology.orchestra.run.vm01.stdout:} 2026-03-06T09:59:08.054 DEBUG:teuthology.orchestra.run.vm01:> sudo mkdir -p /etc/ceph 2026-03-06T09:59:08.090 DEBUG:teuthology.orchestra.run.vm07:> sudo mkdir -p /etc/ceph 2026-03-06T09:59:08.124 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod 777 /etc/ceph 2026-03-06T09:59:08.167 DEBUG:teuthology.orchestra.run.vm07:> sudo chmod 777 /etc/ceph 2026-03-06T09:59:08.193 INFO:tasks.cephadm:Writing seed config... 2026-03-06T09:59:08.193 INFO:tasks.cephadm: override: [mgr] debug mgr = 20 2026-03-06T09:59:08.194 INFO:tasks.cephadm: override: [mgr] debug ms = 1 2026-03-06T09:59:08.194 INFO:tasks.cephadm: override: [mon] debug mon = 20 2026-03-06T09:59:08.194 INFO:tasks.cephadm: override: [mon] debug ms = 1 2026-03-06T09:59:08.194 INFO:tasks.cephadm: override: [mon] debug paxos = 20 2026-03-06T09:59:08.194 INFO:tasks.cephadm: override: [osd] debug ms = 1 2026-03-06T09:59:08.194 INFO:tasks.cephadm: override: [osd] debug osd = 20 2026-03-06T09:59:08.194 INFO:tasks.cephadm: override: [osd] osd mclock iops capacity threshold hdd = 49000 2026-03-06T09:59:08.194 INFO:tasks.cephadm: override: [osd] osd shutdown pgref assert = True 2026-03-06T09:59:08.194 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T09:59:08.194 DEBUG:teuthology.orchestra.run.vm01:> dd of=/home/ubuntu/cephtest/seed.ceph.conf 2026-03-06T09:59:08.226 DEBUG:tasks.cephadm:Final config: [global] # make logging friendly to teuthology log_to_file = true log_to_stderr = false log to journald = false mon cluster log to file = true mon cluster log file level = debug mon clock drift allowed = 1.000 # replicate across OSDs, not hosts osd crush chooseleaf type = 0 #osd pool default size = 2 osd pool default erasure code profile = plugin=jerasure technique=reed_sol_van k=2 m=1 crush-failure-domain=osd # enable some debugging auth debug = true ms die on old message = true ms die on bug = true debug asserts on shutdown = true # adjust warnings mon max pg per osd = 10000# >= luminous mon pg warn max object skew = 0 mon osd allow primary affinity = true mon osd allow pg remap = true mon warn on legacy crush tunables = false mon warn on crush straw calc version zero = false mon warn on no sortbitwise = false mon warn on osd down out interval zero = false mon warn on too few osds = false mon_warn_on_pool_pg_num_not_power_of_two = false # disable pg_autoscaler by default for new pools osd_pool_default_pg_autoscale_mode = off # tests delete pools mon allow pool delete = true fsid = a3ab6672-193a-11f1-b81f-a119763c7190 [osd] osd scrub load threshold = 5.0 osd scrub max interval = 600 osd mclock profile = high_recovery_ops osd recover clone overlap = true osd recovery max chunk = 1048576 osd deep scrub update digest min age = 30 osd map max advance = 10 osd memory target autotune = true # debugging osd debug shutdown = true osd debug op order = true osd debug verify stray on activate = true osd debug pg log writeout = true osd debug verify cached snaps = true osd debug verify missing on start = true osd debug misdirected ops = true osd op queue = debug_random osd op queue cut off = debug_random osd shutdown pgref assert = True bdev debug aio = true osd sloppy crc = true debug ms = 1 debug osd = 20 osd mclock iops capacity threshold hdd = 49000 [mgr] mon reweight min pgs per osd = 4 mon reweight min bytes per osd = 10 mgr/telemetry/nag = false debug mgr = 20 debug ms = 1 [mon] mon data avail warn = 5 mon mgr mkfs grace = 240 mon reweight min pgs per osd = 4 mon osd reporter subtree level = osd mon osd prime pg temp = true mon reweight min bytes per osd = 10 # rotate auth tickets quickly to exercise renewal paths auth mon ticket ttl = 660# 11m auth service ticket ttl = 240# 4m # don't complain about global id reclaim mon_warn_on_insecure_global_id_reclaim = false mon_warn_on_insecure_global_id_reclaim_allowed = false debug mon = 20 debug ms = 1 debug paxos = 20 [client.rgw] rgw cache enabled = true rgw enable ops log = true rgw enable usage log = true 2026-03-06T09:59:08.226 DEBUG:teuthology.orchestra.run.vm01:mon.vm01> sudo journalctl -f -n 0 -u ceph-a3ab6672-193a-11f1-b81f-a119763c7190@mon.vm01.service 2026-03-06T09:59:08.270 INFO:tasks.cephadm:Bootstrapping... 2026-03-06T09:59:08.271 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 -v bootstrap --fsid a3ab6672-193a-11f1-b81f-a119763c7190 --config /home/ubuntu/cephtest/seed.ceph.conf --output-config /etc/ceph/ceph.conf --output-keyring /etc/ceph/ceph.client.admin.keyring --output-pub-ssh-key /home/ubuntu/cephtest/ceph.pub --mon-ip 192.168.123.101 --skip-admin-label && sudo chmod +r /etc/ceph/ceph.client.admin.keyring 2026-03-06T09:59:08.615 INFO:teuthology.orchestra.run.vm01.stdout:-------------------------------------------------------------------------------- 2026-03-06T09:59:08.615 INFO:teuthology.orchestra.run.vm01.stdout:cephadm ['--image', 'harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3', '-v', 'bootstrap', '--fsid', 'a3ab6672-193a-11f1-b81f-a119763c7190', '--config', '/home/ubuntu/cephtest/seed.ceph.conf', '--output-config', '/etc/ceph/ceph.conf', '--output-keyring', '/etc/ceph/ceph.client.admin.keyring', '--output-pub-ssh-key', '/home/ubuntu/cephtest/ceph.pub', '--mon-ip', '192.168.123.101', '--skip-admin-label'] 2026-03-06T09:59:08.615 INFO:teuthology.orchestra.run.vm01.stderr:Specifying an fsid for your cluster offers no advantages and may increase the likelihood of fsid conflicts. 2026-03-06T09:59:08.615 INFO:teuthology.orchestra.run.vm01.stdout:Verifying podman|docker is present... 2026-03-06T09:59:08.647 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stdout 5.8.0 2026-03-06T09:59:08.647 INFO:teuthology.orchestra.run.vm01.stdout:Verifying lvm2 is present... 2026-03-06T09:59:08.647 INFO:teuthology.orchestra.run.vm01.stdout:Verifying time synchronization is in place... 2026-03-06T09:59:08.656 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-06T09:59:08.657 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-06T09:59:08.664 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-06T09:59:08.664 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout inactive 2026-03-06T09:59:08.672 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout enabled 2026-03-06T09:59:08.679 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout active 2026-03-06T09:59:08.679 INFO:teuthology.orchestra.run.vm01.stdout:Unit chronyd.service is enabled and running 2026-03-06T09:59:08.679 INFO:teuthology.orchestra.run.vm01.stdout:Repeating the final host check... 2026-03-06T09:59:08.714 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stdout 5.8.0 2026-03-06T09:59:08.714 INFO:teuthology.orchestra.run.vm01.stdout:podman (/bin/podman) version 5.8.0 is present 2026-03-06T09:59:08.714 INFO:teuthology.orchestra.run.vm01.stdout:systemctl is present 2026-03-06T09:59:08.714 INFO:teuthology.orchestra.run.vm01.stdout:lvcreate is present 2026-03-06T09:59:08.718 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-06T09:59:08.719 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-06T09:59:08.726 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-06T09:59:08.727 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout inactive 2026-03-06T09:59:08.736 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout enabled 2026-03-06T09:59:08.742 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout active 2026-03-06T09:59:08.742 INFO:teuthology.orchestra.run.vm01.stdout:Unit chronyd.service is enabled and running 2026-03-06T09:59:08.742 INFO:teuthology.orchestra.run.vm01.stdout:Host looks OK 2026-03-06T09:59:08.742 INFO:teuthology.orchestra.run.vm01.stdout:Cluster fsid: a3ab6672-193a-11f1-b81f-a119763c7190 2026-03-06T09:59:08.742 INFO:teuthology.orchestra.run.vm01.stdout:Acquiring lock 139891433262672 on /run/cephadm/a3ab6672-193a-11f1-b81f-a119763c7190.lock 2026-03-06T09:59:08.742 INFO:teuthology.orchestra.run.vm01.stdout:Lock 139891433262672 acquired on /run/cephadm/a3ab6672-193a-11f1-b81f-a119763c7190.lock 2026-03-06T09:59:08.742 INFO:teuthology.orchestra.run.vm01.stdout:Verifying IP 192.168.123.101 port 3300 ... 2026-03-06T09:59:08.743 INFO:teuthology.orchestra.run.vm01.stdout:Verifying IP 192.168.123.101 port 6789 ... 2026-03-06T09:59:08.743 INFO:teuthology.orchestra.run.vm01.stdout:Base mon IP(s) is [192.168.123.101:3300, 192.168.123.101:6789], mon addrv is [v2:192.168.123.101:3300,v1:192.168.123.101:6789] 2026-03-06T09:59:08.746 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout default via 192.168.123.1 dev eth0 proto dhcp src 192.168.123.101 metric 100 2026-03-06T09:59:08.746 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout 192.168.123.0/24 dev eth0 proto kernel scope link src 192.168.123.101 metric 100 2026-03-06T09:59:08.749 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout ::1 dev lo proto kernel metric 256 pref medium 2026-03-06T09:59:08.749 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout fe80::/64 dev eth0 proto kernel metric 1024 pref medium 2026-03-06T09:59:08.752 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout 1: lo: mtu 65536 state UNKNOWN qlen 1000 2026-03-06T09:59:08.752 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout inet6 ::1/128 scope host 2026-03-06T09:59:08.752 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-06T09:59:08.752 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout 2: eth0: mtu 1500 state UP qlen 1000 2026-03-06T09:59:08.752 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout inet6 fe80::5055:ff:fe00:1/64 scope link noprefixroute 2026-03-06T09:59:08.752 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-06T09:59:08.752 INFO:teuthology.orchestra.run.vm01.stdout:Mon IP `192.168.123.101` is in CIDR network `192.168.123.0/24` 2026-03-06T09:59:08.752 INFO:teuthology.orchestra.run.vm01.stdout:Mon IP `192.168.123.101` is in CIDR network `192.168.123.0/24` 2026-03-06T09:59:08.752 INFO:teuthology.orchestra.run.vm01.stdout:Inferred mon public CIDR from local network configuration ['192.168.123.0/24', '192.168.123.0/24'] 2026-03-06T09:59:08.753 INFO:teuthology.orchestra.run.vm01.stdout:Internal network (--cluster-network) has not been provided, OSD replication will default to the public_network 2026-03-06T09:59:08.753 INFO:teuthology.orchestra.run.vm01.stdout:Pulling container image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3... 2026-03-06T09:59:09.354 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stdout 306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 2026-03-06T09:59:09.354 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Trying to pull harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3... 2026-03-06T09:59:09.354 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Getting image source signatures 2026-03-06T09:59:09.354 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Copying blob sha256:d21d4233fd3d4dd2f376e5ef084c47891c860682c1de15a9c0357cea5defbc91 2026-03-06T09:59:09.354 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Copying config sha256:306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 2026-03-06T09:59:09.354 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Writing manifest to image destination 2026-03-06T09:59:09.787 INFO:teuthology.orchestra.run.vm01.stdout:ceph: stdout ceph version 19.2.3-47-gc24117fd552 (c24117fd5525679b799527bc1bd1f1dd0a2db5e2) squid (stable) 2026-03-06T09:59:09.787 INFO:teuthology.orchestra.run.vm01.stdout:Ceph version: ceph version 19.2.3-47-gc24117fd552 (c24117fd5525679b799527bc1bd1f1dd0a2db5e2) squid (stable) 2026-03-06T09:59:09.787 INFO:teuthology.orchestra.run.vm01.stdout:Extracting ceph user uid/gid from container image... 2026-03-06T09:59:10.025 INFO:teuthology.orchestra.run.vm01.stdout:stat: stdout 167 167 2026-03-06T09:59:10.025 INFO:teuthology.orchestra.run.vm01.stdout:Creating initial keys... 2026-03-06T09:59:10.265 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-authtool: stdout AQBel6ppFSpDCBAAtZ5TePe0lP6TjQolo81eyg== 2026-03-06T09:59:10.535 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-authtool: stdout AQBel6ppuJrcFhAAYmhhNHjdjRDTYfbPohq90w== 2026-03-06T09:59:10.800 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-authtool: stdout AQBel6ppR0LWJhAAhm8SEelP0S6J3RUFA9iM4g== 2026-03-06T09:59:10.800 INFO:teuthology.orchestra.run.vm01.stdout:Creating initial monmap... 2026-03-06T09:59:11.112 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-06T09:59:11.113 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: stdout setting min_mon_release = quincy 2026-03-06T09:59:11.113 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: set fsid to a3ab6672-193a-11f1-b81f-a119763c7190 2026-03-06T09:59:11.113 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-06T09:59:11.113 INFO:teuthology.orchestra.run.vm01.stdout:monmaptool for vm01 [v2:192.168.123.101:3300,v1:192.168.123.101:6789] on /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-06T09:59:11.113 INFO:teuthology.orchestra.run.vm01.stdout:setting min_mon_release = quincy 2026-03-06T09:59:11.113 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: set fsid to a3ab6672-193a-11f1-b81f-a119763c7190 2026-03-06T09:59:11.113 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-06T09:59:11.113 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T09:59:11.113 INFO:teuthology.orchestra.run.vm01.stdout:Creating mon... 2026-03-06T09:59:11.373 INFO:teuthology.orchestra.run.vm01.stdout:create mon.vm01 on 2026-03-06T09:59:11.739 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /etc/systemd/system/ceph.target. 2026-03-06T09:59:11.922 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph-a3ab6672-193a-11f1-b81f-a119763c7190.target → /etc/systemd/system/ceph-a3ab6672-193a-11f1-b81f-a119763c7190.target. 2026-03-06T09:59:11.922 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph.target.wants/ceph-a3ab6672-193a-11f1-b81f-a119763c7190.target → /etc/systemd/system/ceph-a3ab6672-193a-11f1-b81f-a119763c7190.target. 2026-03-06T09:59:12.107 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-a3ab6672-193a-11f1-b81f-a119763c7190@mon.vm01 2026-03-06T09:59:12.107 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Failed to reset failed state of unit ceph-a3ab6672-193a-11f1-b81f-a119763c7190@mon.vm01.service: Unit ceph-a3ab6672-193a-11f1-b81f-a119763c7190@mon.vm01.service not loaded. 2026-03-06T09:59:12.275 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-a3ab6672-193a-11f1-b81f-a119763c7190.target.wants/ceph-a3ab6672-193a-11f1-b81f-a119763c7190@mon.vm01.service → /etc/systemd/system/ceph-a3ab6672-193a-11f1-b81f-a119763c7190@.service. 2026-03-06T09:59:13.142 INFO:teuthology.orchestra.run.vm01.stdout:firewalld does not appear to be present 2026-03-06T09:59:13.142 INFO:teuthology.orchestra.run.vm01.stdout:Not possible to enable service . firewalld.service is not available 2026-03-06T09:59:13.142 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mon to start... 2026-03-06T09:59:13.143 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mon... 2026-03-06T09:59:13.695 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout cluster: 2026-03-06T09:59:13.695 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout id: a3ab6672-193a-11f1-b81f-a119763c7190 2026-03-06T09:59:13.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout health: HEALTH_OK 2026-03-06T09:59:13.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-06T09:59:13.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout services: 2026-03-06T09:59:13.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon: 1 daemons, quorum vm01 (age 0.32806s) 2026-03-06T09:59:13.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mgr: no daemons active 2026-03-06T09:59:13.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd: 0 osds: 0 up, 0 in 2026-03-06T09:59:13.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-06T09:59:13.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout data: 2026-03-06T09:59:13.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout pools: 0 pools, 0 pgs 2026-03-06T09:59:13.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout objects: 0 objects, 0 B 2026-03-06T09:59:13.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout usage: 0 B used, 0 B / 0 B avail 2026-03-06T09:59:13.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout pgs: 2026-03-06T09:59:13.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-06T09:59:13.696 INFO:teuthology.orchestra.run.vm01.stdout:mon is available 2026-03-06T09:59:13.696 INFO:teuthology.orchestra.run.vm01.stdout:Assimilating anything we can from ceph.conf... 2026-03-06T09:59:14.240 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-06T09:59:14.240 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [global] 2026-03-06T09:59:14.240 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout fsid = a3ab6672-193a-11f1-b81f-a119763c7190 2026-03-06T09:59:14.240 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-06T09:59:14.240 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.101:3300,v1:192.168.123.101:6789] 2026-03-06T09:59:14.240 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-06T09:59:14.240 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-06T09:59:14.240 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-06T09:59:14.240 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-06T09:59:14.240 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-06T09:59:14.240 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-06T09:59:14.240 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-06T09:59:14.240 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-06T09:59:14.240 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [osd] 2026-03-06T09:59:14.240 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-06T09:59:14.241 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-06T09:59:14.241 INFO:teuthology.orchestra.run.vm01.stdout:Generating new minimal ceph.conf... 2026-03-06T09:59:14.675 INFO:teuthology.orchestra.run.vm01.stdout:Restarting the monitor... 2026-03-06T09:59:15.363 INFO:teuthology.orchestra.run.vm01.stdout:Setting public_network to 192.168.123.0/24 in mon config section 2026-03-06T09:59:15.839 INFO:teuthology.orchestra.run.vm01.stdout:Wrote config to /etc/ceph/ceph.conf 2026-03-06T09:59:15.840 INFO:teuthology.orchestra.run.vm01.stdout:Wrote keyring to /etc/ceph/ceph.client.admin.keyring 2026-03-06T09:59:15.840 INFO:teuthology.orchestra.run.vm01.stdout:Creating mgr... 2026-03-06T09:59:15.840 INFO:teuthology.orchestra.run.vm01.stdout:Verifying port 0.0.0.0:9283 ... 2026-03-06T09:59:15.841 INFO:teuthology.orchestra.run.vm01.stdout:Verifying port 0.0.0.0:8765 ... 2026-03-06T09:59:15.841 INFO:teuthology.orchestra.run.vm01.stdout:Verifying port 0.0.0.0:8443 ... 2026-03-06T09:59:16.042 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-a3ab6672-193a-11f1-b81f-a119763c7190@mgr.vm01.ifwqbh 2026-03-06T09:59:16.042 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Failed to reset failed state of unit ceph-a3ab6672-193a-11f1-b81f-a119763c7190@mgr.vm01.ifwqbh.service: Unit ceph-a3ab6672-193a-11f1-b81f-a119763c7190@mgr.vm01.ifwqbh.service not loaded. 2026-03-06T09:59:16.214 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-a3ab6672-193a-11f1-b81f-a119763c7190.target.wants/ceph-a3ab6672-193a-11f1-b81f-a119763c7190@mgr.vm01.ifwqbh.service → /etc/systemd/system/ceph-a3ab6672-193a-11f1-b81f-a119763c7190@.service. 2026-03-06T09:59:16.442 INFO:teuthology.orchestra.run.vm01.stdout:firewalld does not appear to be present 2026-03-06T09:59:16.442 INFO:teuthology.orchestra.run.vm01.stdout:Not possible to enable service . firewalld.service is not available 2026-03-06T09:59:16.442 INFO:teuthology.orchestra.run.vm01.stdout:firewalld does not appear to be present 2026-03-06T09:59:16.442 INFO:teuthology.orchestra.run.vm01.stdout:Not possible to open ports <[9283, 8765, 8443]>. firewalld.service is not available 2026-03-06T09:59:16.442 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mgr to start... 2026-03-06T09:59:16.442 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mgr... 2026-03-06T09:59:17.014 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-06T09:59:17.014 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-06T09:59:17.014 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsid": "a3ab6672-193a-11f1-b81f-a119763c7190", 2026-03-06T09:59:17.015 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T09:59:17.015 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T09:59:17.015 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T09:59:17.015 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T09:59:17.015 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:17.015 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T09:59:17.015 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T09:59:17.015 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 0 2026-03-06T09:59:17.015 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-06T09:59:17.015 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T09:59:17.015 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "vm01" 2026-03-06T09:59:17.015 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-06T09:59:17.015 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_age": 1, 2026-03-06T09:59:17.015 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T09:59:17.015 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T09:59:17.015 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T09:59:17.015 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T09:59:17.015 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:17.016 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T09:59:17.016 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T09:59:17.016 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T09:59:17.016 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T09:59:17.016 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T09:59:17.016 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T09:59:17.016 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T09:59:17.016 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T08:59:13:185373+0000", 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T08:59:13.187095+0000", 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-06T09:59:17.017 INFO:teuthology.orchestra.run.vm01.stdout:mgr not available, waiting (1/15)... 2026-03-06T09:59:19.751 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-06T09:59:19.752 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-06T09:59:19.752 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsid": "a3ab6672-193a-11f1-b81f-a119763c7190", 2026-03-06T09:59:19.752 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T09:59:19.752 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T09:59:19.752 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T09:59:19.752 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T09:59:19.752 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:19.752 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T09:59:19.752 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T09:59:19.752 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 0 2026-03-06T09:59:19.752 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-06T09:59:19.752 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T09:59:19.752 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "vm01" 2026-03-06T09:59:19.752 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-06T09:59:19.752 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_age": 4, 2026-03-06T09:59:19.752 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T09:59:19.752 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T09:59:19.752 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T09:59:19.752 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T09:59:19.752 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:19.752 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T09:59:19.752 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T09:59:19.752 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T09:59:19.752 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T09:59:19.752 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T09:59:19.752 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T09:59:19.753 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T09:59:19.753 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T08:59:13:185373+0000", 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T08:59:13.187095+0000", 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-06T09:59:19.754 INFO:teuthology.orchestra.run.vm01.stdout:mgr not available, waiting (2/15)... 2026-03-06T09:59:22.331 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-06T09:59:22.331 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-06T09:59:22.332 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsid": "a3ab6672-193a-11f1-b81f-a119763c7190", 2026-03-06T09:59:22.332 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T09:59:22.332 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T09:59:22.332 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T09:59:22.332 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T09:59:22.332 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:22.332 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T09:59:22.332 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T09:59:22.332 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 0 2026-03-06T09:59:22.332 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-06T09:59:22.332 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T09:59:22.332 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "vm01" 2026-03-06T09:59:22.332 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-06T09:59:22.332 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_age": 6, 2026-03-06T09:59:22.332 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T09:59:22.332 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T09:59:22.332 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T09:59:22.332 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T09:59:22.332 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:22.332 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T09:59:22.332 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T09:59:22.332 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T09:59:22.332 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T09:59:22.332 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T09:59:22.332 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T08:59:13:185373+0000", 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T08:59:13.187095+0000", 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-06T09:59:22.333 INFO:teuthology.orchestra.run.vm01.stdout:mgr not available, waiting (3/15)... 2026-03-06T09:59:24.938 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-06T09:59:24.938 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-06T09:59:24.938 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsid": "a3ab6672-193a-11f1-b81f-a119763c7190", 2026-03-06T09:59:24.938 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T09:59:24.938 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T09:59:24.938 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T09:59:24.938 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T09:59:24.938 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:24.938 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T09:59:24.938 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T09:59:24.938 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 0 2026-03-06T09:59:24.938 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-06T09:59:24.938 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T09:59:24.938 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "vm01" 2026-03-06T09:59:24.938 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-06T09:59:24.938 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_age": 9, 2026-03-06T09:59:24.939 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T09:59:24.939 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T09:59:24.939 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T09:59:24.939 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T09:59:24.939 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:24.939 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T09:59:24.939 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T09:59:24.939 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T09:59:24.939 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T09:59:24.939 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T08:59:13:185373+0000", 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T08:59:13.187095+0000", 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-06T09:59:24.940 INFO:teuthology.orchestra.run.vm01.stdout:mgr not available, waiting (4/15)... 2026-03-06T09:59:27.460 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-06T09:59:27.460 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-06T09:59:27.461 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsid": "a3ab6672-193a-11f1-b81f-a119763c7190", 2026-03-06T09:59:27.461 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T09:59:27.461 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T09:59:27.461 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T09:59:27.461 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T09:59:27.461 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:27.461 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T09:59:27.461 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T09:59:27.461 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 0 2026-03-06T09:59:27.461 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-06T09:59:27.461 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T09:59:27.461 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "vm01" 2026-03-06T09:59:27.461 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-06T09:59:27.461 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_age": 11, 2026-03-06T09:59:27.461 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T09:59:27.461 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T08:59:13:185373+0000", 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T08:59:13.187095+0000", 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-06T09:59:27.462 INFO:teuthology.orchestra.run.vm01.stdout:mgr not available, waiting (5/15)... 2026-03-06T09:59:30.042 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-06T09:59:30.043 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsid": "a3ab6672-193a-11f1-b81f-a119763c7190", 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 0 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "vm01" 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_age": 14, 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T08:59:13:185373+0000", 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T09:59:30.044 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T09:59:30.045 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-06T09:59:30.045 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T09:59:30.045 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:30.045 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T09:59:30.045 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T09:59:30.045 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T08:59:13.187095+0000", 2026-03-06T09:59:30.045 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T09:59:30.045 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-06T09:59:30.045 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T09:59:30.045 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-06T09:59:30.045 INFO:teuthology.orchestra.run.vm01.stdout:mgr is available 2026-03-06T09:59:30.582 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-06T09:59:30.582 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [global] 2026-03-06T09:59:30.582 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout fsid = a3ab6672-193a-11f1-b81f-a119763c7190 2026-03-06T09:59:30.582 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-06T09:59:30.582 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.101:3300,v1:192.168.123.101:6789] 2026-03-06T09:59:30.582 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-06T09:59:30.582 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-06T09:59:30.582 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-06T09:59:30.583 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-06T09:59:30.583 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-06T09:59:30.583 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-06T09:59:30.583 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-06T09:59:30.583 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-06T09:59:30.583 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [osd] 2026-03-06T09:59:30.583 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-06T09:59:30.583 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-06T09:59:30.583 INFO:teuthology.orchestra.run.vm01.stdout:Enabling cephadm module... 2026-03-06T09:59:31.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-06T09:59:31.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 5, 2026-03-06T09:59:31.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-06T09:59:31.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "active_name": "vm01.ifwqbh", 2026-03-06T09:59:31.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-06T09:59:31.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-06T09:59:31.984 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for the mgr to restart... 2026-03-06T09:59:31.985 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mgr epoch 5... 2026-03-06T09:59:43.912 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-06T09:59:43.912 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 7, 2026-03-06T09:59:43.912 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-06T09:59:43.912 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-06T09:59:43.912 INFO:teuthology.orchestra.run.vm01.stdout:mgr epoch 5 is available 2026-03-06T09:59:43.912 INFO:teuthology.orchestra.run.vm01.stdout:Setting orchestrator backend to cephadm... 2026-03-06T09:59:44.918 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout value unchanged 2026-03-06T09:59:44.918 INFO:teuthology.orchestra.run.vm01.stdout:Generating ssh key... 2026-03-06T09:59:45.957 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDC0Sm9uaNShDaKx1Xbm1i4JVhzGup9qTf0vQcVRc7GvbQpkh0np3AR1FFj+vdjbVvgbne4lBlzduu2PjjnbnbwagiQuvfVRpUguYB91nFb/ZEkOFwWNwrLCtuNQqDwD8mvlvptgB9sb+Ok2ckUtthYxVkrMwI0cDY7lZ/K++IRud97MMTpDu/LFwJ47I89AP1mWsakQwsdD98AcqxHvPOT9id8dSRLocyvaAinK08//jbMRDYJtR8FqECUeKqi+6MDzNNAXpWgGP8pKSMnXiACk2Xa+LIqbXCh91X55ctMb5T0esk7TCG/NWrJ85j9MIHEoonbWuEod6MqycJsn3tbEgE4MWhdMjvIIuJltd65gIYHbgKDEG1q7ppw9LS+75ZR2xiD+NOBLaTHpJQdN9qrSb8BNbtFeFuwr5M5WmmI4myisaC4i+lxcbJTWxxvb2syJ0wYqAfQBJuzG9BHqiDMH4AkMN0Yjn90UWrCIx3etQBBTnKGuD11vD1V4+4KvDs= ceph-a3ab6672-193a-11f1-b81f-a119763c7190 2026-03-06T09:59:45.957 INFO:teuthology.orchestra.run.vm01.stdout:Wrote public SSH key to /home/ubuntu/cephtest/ceph.pub 2026-03-06T09:59:45.957 INFO:teuthology.orchestra.run.vm01.stdout:Adding key to root@localhost authorized_keys... 2026-03-06T09:59:45.958 INFO:teuthology.orchestra.run.vm01.stdout:Adding host vm01... 2026-03-06T09:59:48.995 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Added host 'vm01' with addr '192.168.123.101' 2026-03-06T09:59:48.995 INFO:teuthology.orchestra.run.vm01.stdout:Deploying mon service with default placement... 2026-03-06T09:59:49.548 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Scheduled mon update... 2026-03-06T09:59:49.549 INFO:teuthology.orchestra.run.vm01.stdout:Deploying mgr service with default placement... 2026-03-06T09:59:50.079 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Scheduled mgr update... 2026-03-06T09:59:50.079 INFO:teuthology.orchestra.run.vm01.stdout:Deploying crash service with default placement... 2026-03-06T09:59:50.590 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Scheduled crash update... 2026-03-06T09:59:50.590 INFO:teuthology.orchestra.run.vm01.stdout:Deploying ceph-exporter service with default placement... 2026-03-06T09:59:51.177 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Scheduled ceph-exporter update... 2026-03-06T09:59:51.177 INFO:teuthology.orchestra.run.vm01.stdout:Deploying prometheus service with default placement... 2026-03-06T09:59:51.704 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Scheduled prometheus update... 2026-03-06T09:59:51.704 INFO:teuthology.orchestra.run.vm01.stdout:Deploying grafana service with default placement... 2026-03-06T09:59:52.260 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Scheduled grafana update... 2026-03-06T09:59:52.260 INFO:teuthology.orchestra.run.vm01.stdout:Deploying node-exporter service with default placement... 2026-03-06T09:59:52.794 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Scheduled node-exporter update... 2026-03-06T09:59:52.794 INFO:teuthology.orchestra.run.vm01.stdout:Deploying alertmanager service with default placement... 2026-03-06T09:59:53.359 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Scheduled alertmanager update... 2026-03-06T09:59:54.368 INFO:teuthology.orchestra.run.vm01.stdout:Enabling the dashboard module... 2026-03-06T09:59:55.932 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-06T09:59:55.932 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 9, 2026-03-06T09:59:55.932 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-06T09:59:55.932 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "active_name": "vm01.ifwqbh", 2026-03-06T09:59:55.932 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-06T09:59:55.932 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-06T09:59:55.932 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for the mgr to restart... 2026-03-06T09:59:55.932 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mgr epoch 9... 2026-03-06T09:59:56.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 09:59:56 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/2958663641' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-06T10:00:06.239 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:06 vm01 ceph-mon[50931]: Active manager daemon vm01.ifwqbh restarted 2026-03-06T10:00:06.240 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:06 vm01 ceph-mon[50931]: Activating manager daemon vm01.ifwqbh 2026-03-06T10:00:06.240 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:06 vm01 ceph-mon[50931]: osdmap e3: 0 total, 0 up, 0 in 2026-03-06T10:00:06.240 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:06 vm01 ceph-mon[50931]: mgrmap e10: vm01.ifwqbh(active, starting, since 0.00679361s) 2026-03-06T10:00:06.240 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:06 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-06T10:00:06.240 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:06 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mgr metadata", "who": "vm01.ifwqbh", "id": "vm01.ifwqbh"}]: dispatch 2026-03-06T10:00:06.240 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:06 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-06T10:00:06.240 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:06 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-06T10:00:06.240 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:06 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-06T10:00:06.240 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:06 vm01 ceph-mon[50931]: Manager daemon vm01.ifwqbh is now available 2026-03-06T10:00:06.240 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:06 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.ifwqbh/mirror_snapshot_schedule"}]: dispatch 2026-03-06T10:00:06.240 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:06 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T10:00:07.277 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-06T10:00:07.277 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 11, 2026-03-06T10:00:07.277 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-06T10:00:07.277 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-06T10:00:07.277 INFO:teuthology.orchestra.run.vm01.stdout:mgr epoch 9 is available 2026-03-06T10:00:07.277 INFO:teuthology.orchestra.run.vm01.stdout:Generating a dashboard self-signed certificate... 2026-03-06T10:00:07.373 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:07 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.ifwqbh/trash_purge_schedule"}]: dispatch 2026-03-06T10:00:07.373 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:07 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:07.373 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:07 vm01 ceph-mon[50931]: mgrmap e11: vm01.ifwqbh(active, since 1.00991s) 2026-03-06T10:00:07.897 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Self-signed certificate created 2026-03-06T10:00:07.897 INFO:teuthology.orchestra.run.vm01.stdout:Creating initial admin user... 2026-03-06T10:00:08.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:08 vm01 ceph-mon[50931]: [06/Mar/2026:09:00:07] ENGINE Bus STARTING 2026-03-06T10:00:08.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:08 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:08.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:08 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:08.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:08 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:08.604 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout {"username": "admin", "password": "$2b$12$10yMGMVnKtI6cqSbc0fMau4dzoBbmG8dl6YpplmTo2XKy6lhtRw.G", "roles": ["administrator"], "name": null, "email": null, "lastUpdate": 1772787608, "enabled": true, "pwdExpirationDate": null, "pwdUpdateRequired": true} 2026-03-06T10:00:08.604 INFO:teuthology.orchestra.run.vm01.stdout:Fetching dashboard port number... 2026-03-06T10:00:09.080 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 8443 2026-03-06T10:00:09.080 INFO:teuthology.orchestra.run.vm01.stdout:firewalld does not appear to be present 2026-03-06T10:00:09.080 INFO:teuthology.orchestra.run.vm01.stdout:Not possible to open ports <[8443]>. firewalld.service is not available 2026-03-06T10:00:09.080 INFO:teuthology.orchestra.run.vm01.stdout:Ceph Dashboard is now available at: 2026-03-06T10:00:09.080 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:00:09.080 INFO:teuthology.orchestra.run.vm01.stdout: URL: https://vm01.local:8443/ 2026-03-06T10:00:09.080 INFO:teuthology.orchestra.run.vm01.stdout: User: admin 2026-03-06T10:00:09.080 INFO:teuthology.orchestra.run.vm01.stdout: Password: yyc97w5wn0 2026-03-06T10:00:09.080 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:00:09.081 INFO:teuthology.orchestra.run.vm01.stdout:Saving cluster configuration to /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/config directory 2026-03-06T10:00:09.201 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:09 vm01 ceph-mon[50931]: [06/Mar/2026:09:00:07] ENGINE Serving on https://192.168.123.101:7150 2026-03-06T10:00:09.201 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:09 vm01 ceph-mon[50931]: [06/Mar/2026:09:00:07] ENGINE Client ('192.168.123.101', 57738) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-06T10:00:09.201 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:09 vm01 ceph-mon[50931]: [06/Mar/2026:09:00:07] ENGINE Serving on http://192.168.123.101:8765 2026-03-06T10:00:09.201 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:09 vm01 ceph-mon[50931]: [06/Mar/2026:09:00:07] ENGINE Bus STARTED 2026-03-06T10:00:09.201 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:09 vm01 ceph-mon[50931]: from='client.14180 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:00:09.201 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:09 vm01 ceph-mon[50931]: from='client.14182 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:00:09.201 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:09 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:09.201 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:09 vm01 ceph-mon[50931]: mgrmap e12: vm01.ifwqbh(active, since 2s) 2026-03-06T10:00:09.201 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:09 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/2707147569' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-06T10:00:09.681 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stderr set mgr/dashboard/cluster/status 2026-03-06T10:00:09.681 INFO:teuthology.orchestra.run.vm01.stdout:You can access the Ceph CLI as following in case of multi-cluster or non-default config: 2026-03-06T10:00:09.681 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:00:09.681 INFO:teuthology.orchestra.run.vm01.stdout: sudo /home/ubuntu/cephtest/cephadm shell --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring 2026-03-06T10:00:09.681 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:00:09.681 INFO:teuthology.orchestra.run.vm01.stdout:Or, if you are only running a single cluster on this host: 2026-03-06T10:00:09.681 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:00:09.681 INFO:teuthology.orchestra.run.vm01.stdout: sudo /home/ubuntu/cephtest/cephadm shell 2026-03-06T10:00:09.681 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:00:09.681 INFO:teuthology.orchestra.run.vm01.stdout:Please consider enabling telemetry to help improve Ceph: 2026-03-06T10:00:09.681 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:00:09.681 INFO:teuthology.orchestra.run.vm01.stdout: ceph telemetry on 2026-03-06T10:00:09.681 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:00:09.681 INFO:teuthology.orchestra.run.vm01.stdout:For more information see: 2026-03-06T10:00:09.681 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:00:09.681 INFO:teuthology.orchestra.run.vm01.stdout: https://docs.ceph.com/en/latest/mgr/telemetry/ 2026-03-06T10:00:09.681 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:00:09.681 INFO:teuthology.orchestra.run.vm01.stdout:Bootstrap complete. 2026-03-06T10:00:09.720 INFO:tasks.cephadm:Fetching config... 2026-03-06T10:00:09.720 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T10:00:09.720 DEBUG:teuthology.orchestra.run.vm01:> dd if=/etc/ceph/ceph.conf of=/dev/stdout 2026-03-06T10:00:09.740 INFO:tasks.cephadm:Fetching client.admin keyring... 2026-03-06T10:00:09.740 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T10:00:09.740 DEBUG:teuthology.orchestra.run.vm01:> dd if=/etc/ceph/ceph.client.admin.keyring of=/dev/stdout 2026-03-06T10:00:09.807 INFO:tasks.cephadm:Fetching mon keyring... 2026-03-06T10:00:09.807 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T10:00:09.807 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/keyring of=/dev/stdout 2026-03-06T10:00:09.881 INFO:tasks.cephadm:Fetching pub ssh key... 2026-03-06T10:00:09.881 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T10:00:09.881 DEBUG:teuthology.orchestra.run.vm01:> dd if=/home/ubuntu/cephtest/ceph.pub of=/dev/stdout 2026-03-06T10:00:09.940 INFO:tasks.cephadm:Installing pub ssh key for root users... 2026-03-06T10:00:09.940 DEBUG:teuthology.orchestra.run.vm01:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDC0Sm9uaNShDaKx1Xbm1i4JVhzGup9qTf0vQcVRc7GvbQpkh0np3AR1FFj+vdjbVvgbne4lBlzduu2PjjnbnbwagiQuvfVRpUguYB91nFb/ZEkOFwWNwrLCtuNQqDwD8mvlvptgB9sb+Ok2ckUtthYxVkrMwI0cDY7lZ/K++IRud97MMTpDu/LFwJ47I89AP1mWsakQwsdD98AcqxHvPOT9id8dSRLocyvaAinK08//jbMRDYJtR8FqECUeKqi+6MDzNNAXpWgGP8pKSMnXiACk2Xa+LIqbXCh91X55ctMb5T0esk7TCG/NWrJ85j9MIHEoonbWuEod6MqycJsn3tbEgE4MWhdMjvIIuJltd65gIYHbgKDEG1q7ppw9LS+75ZR2xiD+NOBLaTHpJQdN9qrSb8BNbtFeFuwr5M5WmmI4myisaC4i+lxcbJTWxxvb2syJ0wYqAfQBJuzG9BHqiDMH4AkMN0Yjn90UWrCIx3etQBBTnKGuD11vD1V4+4KvDs= ceph-a3ab6672-193a-11f1-b81f-a119763c7190' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-06T10:00:10.022 INFO:teuthology.orchestra.run.vm01.stdout:ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDC0Sm9uaNShDaKx1Xbm1i4JVhzGup9qTf0vQcVRc7GvbQpkh0np3AR1FFj+vdjbVvgbne4lBlzduu2PjjnbnbwagiQuvfVRpUguYB91nFb/ZEkOFwWNwrLCtuNQqDwD8mvlvptgB9sb+Ok2ckUtthYxVkrMwI0cDY7lZ/K++IRud97MMTpDu/LFwJ47I89AP1mWsakQwsdD98AcqxHvPOT9id8dSRLocyvaAinK08//jbMRDYJtR8FqECUeKqi+6MDzNNAXpWgGP8pKSMnXiACk2Xa+LIqbXCh91X55ctMb5T0esk7TCG/NWrJ85j9MIHEoonbWuEod6MqycJsn3tbEgE4MWhdMjvIIuJltd65gIYHbgKDEG1q7ppw9LS+75ZR2xiD+NOBLaTHpJQdN9qrSb8BNbtFeFuwr5M5WmmI4myisaC4i+lxcbJTWxxvb2syJ0wYqAfQBJuzG9BHqiDMH4AkMN0Yjn90UWrCIx3etQBBTnKGuD11vD1V4+4KvDs= ceph-a3ab6672-193a-11f1-b81f-a119763c7190 2026-03-06T10:00:10.035 DEBUG:teuthology.orchestra.run.vm07:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDC0Sm9uaNShDaKx1Xbm1i4JVhzGup9qTf0vQcVRc7GvbQpkh0np3AR1FFj+vdjbVvgbne4lBlzduu2PjjnbnbwagiQuvfVRpUguYB91nFb/ZEkOFwWNwrLCtuNQqDwD8mvlvptgB9sb+Ok2ckUtthYxVkrMwI0cDY7lZ/K++IRud97MMTpDu/LFwJ47I89AP1mWsakQwsdD98AcqxHvPOT9id8dSRLocyvaAinK08//jbMRDYJtR8FqECUeKqi+6MDzNNAXpWgGP8pKSMnXiACk2Xa+LIqbXCh91X55ctMb5T0esk7TCG/NWrJ85j9MIHEoonbWuEod6MqycJsn3tbEgE4MWhdMjvIIuJltd65gIYHbgKDEG1q7ppw9LS+75ZR2xiD+NOBLaTHpJQdN9qrSb8BNbtFeFuwr5M5WmmI4myisaC4i+lxcbJTWxxvb2syJ0wYqAfQBJuzG9BHqiDMH4AkMN0Yjn90UWrCIx3etQBBTnKGuD11vD1V4+4KvDs= ceph-a3ab6672-193a-11f1-b81f-a119763c7190' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-06T10:00:10.069 INFO:teuthology.orchestra.run.vm07.stdout:ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDC0Sm9uaNShDaKx1Xbm1i4JVhzGup9qTf0vQcVRc7GvbQpkh0np3AR1FFj+vdjbVvgbne4lBlzduu2PjjnbnbwagiQuvfVRpUguYB91nFb/ZEkOFwWNwrLCtuNQqDwD8mvlvptgB9sb+Ok2ckUtthYxVkrMwI0cDY7lZ/K++IRud97MMTpDu/LFwJ47I89AP1mWsakQwsdD98AcqxHvPOT9id8dSRLocyvaAinK08//jbMRDYJtR8FqECUeKqi+6MDzNNAXpWgGP8pKSMnXiACk2Xa+LIqbXCh91X55ctMb5T0esk7TCG/NWrJ85j9MIHEoonbWuEod6MqycJsn3tbEgE4MWhdMjvIIuJltd65gIYHbgKDEG1q7ppw9LS+75ZR2xiD+NOBLaTHpJQdN9qrSb8BNbtFeFuwr5M5WmmI4myisaC4i+lxcbJTWxxvb2syJ0wYqAfQBJuzG9BHqiDMH4AkMN0Yjn90UWrCIx3etQBBTnKGuD11vD1V4+4KvDs= ceph-a3ab6672-193a-11f1-b81f-a119763c7190 2026-03-06T10:00:10.081 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph config set mgr mgr/cephadm/allow_ptrace true 2026-03-06T10:00:10.479 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:00:10.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:10 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/2942534136' entity='client.admin' 2026-03-06T10:00:11.049 INFO:tasks.cephadm:Distributing conf and client.admin keyring to all hosts + 0755 2026-03-06T10:00:11.049 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph orch client-keyring set client.admin '*' --mode 0755 2026-03-06T10:00:11.578 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:00:12.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:11 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/1673793612' entity='client.admin' 2026-03-06T10:00:12.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:11 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:12.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:11 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:12.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:11 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-06T10:00:12.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:11 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:12.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:11 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm01", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-06T10:00:12.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:11 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' cmd='[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm01", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]': finished 2026-03-06T10:00:12.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:11 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:00:12.325 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm07 2026-03-06T10:00:12.325 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-06T10:00:12.325 DEBUG:teuthology.orchestra.run.vm07:> dd of=/etc/ceph/ceph.conf 2026-03-06T10:00:12.341 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-06T10:00:12.341 DEBUG:teuthology.orchestra.run.vm07:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-06T10:00:12.398 INFO:tasks.cephadm:Adding host vm07 to orchestrator... 2026-03-06T10:00:12.398 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph orch host add vm07 2026-03-06T10:00:12.802 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:00:13.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:13 vm01 ceph-mon[50931]: Deploying daemon ceph-exporter.vm01 on vm01 2026-03-06T10:00:13.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:13 vm01 ceph-mon[50931]: from='client.14190 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:00:13.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:13 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:13.367 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:13 vm01 ceph-mon[50931]: mgrmap e13: vm01.ifwqbh(active, since 6s) 2026-03-06T10:00:14.301 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:14 vm01 ceph-mon[50931]: from='client.14192 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm07", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:00:15.351 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:15 vm01 ceph-mon[50931]: Deploying cephadm binary to vm07 2026-03-06T10:00:15.351 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:15 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:15.351 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:15 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:15.351 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:15 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:15.351 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:15 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:15.351 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:15 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm01", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-06T10:00:15.351 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:15 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' cmd='[{"prefix": "auth get-or-create", "entity": "client.crash.vm01", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]': finished 2026-03-06T10:00:15.351 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:15 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:00:15.969 INFO:teuthology.orchestra.run.vm01.stdout:Added host 'vm07' with addr '192.168.123.107' 2026-03-06T10:00:16.239 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:16 vm01 ceph-mon[50931]: Deploying daemon crash.vm01 on vm01 2026-03-06T10:00:16.239 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:16 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:16.239 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:16 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:16.239 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:16 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:16.239 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:16 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:16.239 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:16 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:16.239 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:16 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:16.269 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph orch host ls --format=json 2026-03-06T10:00:16.722 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:00:17.082 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:00:17.082 INFO:teuthology.orchestra.run.vm01.stdout:[{"addr": "192.168.123.101", "hostname": "vm01", "labels": [], "status": ""}, {"addr": "192.168.123.107", "hostname": "vm07", "labels": [], "status": ""}] 2026-03-06T10:00:17.267 INFO:tasks.cephadm:Setting crush tunables to default 2026-03-06T10:00:17.267 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd crush tunables default 2026-03-06T10:00:17.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:17 vm01 ceph-mon[50931]: Deploying daemon node-exporter.vm01 on vm01 2026-03-06T10:00:17.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:17 vm01 ceph-mon[50931]: Added host vm07 2026-03-06T10:00:17.606 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:00:18.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:18 vm01 ceph-mon[50931]: from='client.14195 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T10:00:19.308 INFO:teuthology.orchestra.run.vm01.stderr:adjusted tunables profile to default 2026-03-06T10:00:19.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:19 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/390609124' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-06T10:00:19.982 INFO:tasks.cephadm:Adding mon.vm01 on vm01 2026-03-06T10:00:19.982 INFO:tasks.cephadm:Adding mon.vm07 on vm07 2026-03-06T10:00:19.982 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph orch apply mon '2;vm01:192.168.123.101=vm01;vm07:192.168.123.107=vm07' 2026-03-06T10:00:20.317 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:20.349 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:20.730 INFO:teuthology.orchestra.run.vm07.stdout:Scheduled mon update... 2026-03-06T10:00:20.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:20 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/390609124' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-06T10:00:20.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:20 vm01 ceph-mon[50931]: osdmap e4: 0 total, 0 up, 0 in 2026-03-06T10:00:20.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:20 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:20.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:20 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:20.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:20 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:20.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:20 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:20.917 DEBUG:teuthology.orchestra.run.vm07:mon.vm07> sudo journalctl -f -n 0 -u ceph-a3ab6672-193a-11f1-b81f-a119763c7190@mon.vm07.service 2026-03-06T10:00:20.919 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-06T10:00:20.919 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph mon dump -f json 2026-03-06T10:00:21.289 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:21.331 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:21.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:21 vm01 ceph-mon[50931]: Deploying daemon alertmanager.vm01 on vm01 2026-03-06T10:00:21.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:21 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:21.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:21 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:21.848 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-06T10:00:21.848 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":1,"fsid":"a3ab6672-193a-11f1-b81f-a119763c7190","modified":"2026-03-06T08:59:10.955430Z","created":"2026-03-06T08:59:10.955430Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-06T10:00:21.848 INFO:teuthology.orchestra.run.vm07.stderr:dumped monmap epoch 1 2026-03-06T10:00:22.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:22 vm01 ceph-mon[50931]: from='client.14199 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "2;vm01:192.168.123.101=vm01;vm07:192.168.123.107=vm07", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:00:22.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:22 vm01 ceph-mon[50931]: Saving service mon spec with placement vm01:192.168.123.101=vm01;vm07:192.168.123.107=vm07;count:2 2026-03-06T10:00:22.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:22 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/3902814092' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T10:00:23.303 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-06T10:00:23.303 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph mon dump -f json 2026-03-06T10:00:23.637 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:23.680 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:24.084 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-06T10:00:24.085 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":1,"fsid":"a3ab6672-193a-11f1-b81f-a119763c7190","modified":"2026-03-06T08:59:10.955430Z","created":"2026-03-06T08:59:10.955430Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-06T10:00:24.085 INFO:teuthology.orchestra.run.vm07.stderr:dumped monmap epoch 1 2026-03-06T10:00:24.409 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:24 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/3825347213' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T10:00:24.409 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:24 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:25.273 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-06T10:00:25.273 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph mon dump -f json 2026-03-06T10:00:25.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:25 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:25.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:25 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:25.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:25 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:25.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:25 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:25.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:25 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:25.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:25 vm01 ceph-mon[50931]: Regenerating cephadm self-signed grafana TLS certificates 2026-03-06T10:00:25.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:25 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:25.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:25 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:25.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:25 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "dashboard set-grafana-api-ssl-verify", "value": "false"}]: dispatch 2026-03-06T10:00:25.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:25 vm01 ceph-mon[50931]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-ssl-verify", "value": "false"}]: dispatch 2026-03-06T10:00:25.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:25 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:25.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:25 vm01 ceph-mon[50931]: Deploying daemon grafana.vm01 on vm01 2026-03-06T10:00:25.629 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:25.677 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:26.073 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-06T10:00:26.073 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":1,"fsid":"a3ab6672-193a-11f1-b81f-a119763c7190","modified":"2026-03-06T08:59:10.955430Z","created":"2026-03-06T08:59:10.955430Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-06T10:00:26.073 INFO:teuthology.orchestra.run.vm07.stderr:dumped monmap epoch 1 2026-03-06T10:00:26.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:26 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/2128036592' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T10:00:27.224 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-06T10:00:27.224 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph mon dump -f json 2026-03-06T10:00:27.568 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:27.614 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:27.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:27 vm01 ceph-mon[50931]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:00:27.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:27 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:28.017 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-06T10:00:28.017 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":1,"fsid":"a3ab6672-193a-11f1-b81f-a119763c7190","modified":"2026-03-06T08:59:10.955430Z","created":"2026-03-06T08:59:10.955430Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-06T10:00:28.017 INFO:teuthology.orchestra.run.vm07.stderr:dumped monmap epoch 1 2026-03-06T10:00:28.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:28 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/2561911696' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T10:00:29.193 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-06T10:00:29.193 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph mon dump -f json 2026-03-06T10:00:29.520 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:29.563 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:29.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:29 vm01 ceph-mon[50931]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:00:29.943 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-06T10:00:29.943 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":1,"fsid":"a3ab6672-193a-11f1-b81f-a119763c7190","modified":"2026-03-06T08:59:10.955430Z","created":"2026-03-06T08:59:10.955430Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-06T10:00:29.943 INFO:teuthology.orchestra.run.vm07.stderr:dumped monmap epoch 1 2026-03-06T10:00:30.344 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:30 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/1632896300' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T10:00:31.120 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-06T10:00:31.120 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph mon dump -f json 2026-03-06T10:00:31.456 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:31.486 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:31 vm01 ceph-mon[50931]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:00:31.486 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:31 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:31.486 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:31 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:31.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:31 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:31.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:31 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:31.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:31 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:31.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:31 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:31.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:31 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:31.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:31 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:31.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:31 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:31.498 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:31.885 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-06T10:00:31.885 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":1,"fsid":"a3ab6672-193a-11f1-b81f-a119763c7190","modified":"2026-03-06T08:59:10.955430Z","created":"2026-03-06T08:59:10.955430Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-06T10:00:31.885 INFO:teuthology.orchestra.run.vm07.stderr:dumped monmap epoch 1 2026-03-06T10:00:32.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:32 vm01 ceph-mon[50931]: Deploying daemon prometheus.vm01 on vm01 2026-03-06T10:00:32.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:32 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/2291775442' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T10:00:33.063 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-06T10:00:33.063 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph mon dump -f json 2026-03-06T10:00:33.392 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:33.436 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:33.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:33 vm01 ceph-mon[50931]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:00:33.825 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-06T10:00:33.825 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":1,"fsid":"a3ab6672-193a-11f1-b81f-a119763c7190","modified":"2026-03-06T08:59:10.955430Z","created":"2026-03-06T08:59:10.955430Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-06T10:00:33.825 INFO:teuthology.orchestra.run.vm07.stderr:dumped monmap epoch 1 2026-03-06T10:00:34.420 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:34 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/2988926689' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T10:00:34.420 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:34 vm01 ceph-mon[50931]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:00:35.014 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-06T10:00:35.014 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph mon dump -f json 2026-03-06T10:00:35.351 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:35.389 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:35.789 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-06T10:00:35.789 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":1,"fsid":"a3ab6672-193a-11f1-b81f-a119763c7190","modified":"2026-03-06T08:59:10.955430Z","created":"2026-03-06T08:59:10.955430Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-06T10:00:35.789 INFO:teuthology.orchestra.run.vm07.stderr:dumped monmap epoch 1 2026-03-06T10:00:36.170 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:35 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/3162672634' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T10:00:36.977 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-06T10:00:36.977 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph mon dump -f json 2026-03-06T10:00:37.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:36 vm01 ceph-mon[50931]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:00:37.323 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:37.366 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:37.762 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-06T10:00:37.762 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":1,"fsid":"a3ab6672-193a-11f1-b81f-a119763c7190","modified":"2026-03-06T08:59:10.955430Z","created":"2026-03-06T08:59:10.955430Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-06T10:00:37.762 INFO:teuthology.orchestra.run.vm07.stderr:dumped monmap epoch 1 2026-03-06T10:00:38.246 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:37 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:38.246 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:37 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:38.246 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:37 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:38.246 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:37 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mgr module enable", "module": "prometheus"}]: dispatch 2026-03-06T10:00:38.246 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:37 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/1583220264' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T10:00:38.938 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-06T10:00:38.938 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph mon dump -f json 2026-03-06T10:00:39.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:38 vm01 ceph-mon[50931]: from='mgr.14168 192.168.123.101:0/3242816297' entity='mgr.vm01.ifwqbh' cmd='[{"prefix": "mgr module enable", "module": "prometheus"}]': finished 2026-03-06T10:00:39.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:38 vm01 ceph-mon[50931]: mgrmap e14: vm01.ifwqbh(active, since 31s) 2026-03-06T10:00:39.271 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:39.314 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:39.729 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-06T10:00:39.729 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":1,"fsid":"a3ab6672-193a-11f1-b81f-a119763c7190","modified":"2026-03-06T08:59:10.955430Z","created":"2026-03-06T08:59:10.955430Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-06T10:00:39.729 INFO:teuthology.orchestra.run.vm07.stderr:dumped monmap epoch 1 2026-03-06T10:00:40.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:39 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/2517151779' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T10:00:40.917 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-06T10:00:40.917 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph mon dump -f json 2026-03-06T10:00:41.283 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:41.331 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:41.755 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-06T10:00:41.756 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":1,"fsid":"a3ab6672-193a-11f1-b81f-a119763c7190","modified":"2026-03-06T08:59:10.955430Z","created":"2026-03-06T08:59:10.955430Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-06T10:00:41.756 INFO:teuthology.orchestra.run.vm07.stderr:dumped monmap epoch 1 2026-03-06T10:00:42.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:41 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/2021840421' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T10:00:42.920 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-06T10:00:42.920 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph mon dump -f json 2026-03-06T10:00:43.272 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:43.317 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:43.708 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-06T10:00:43.708 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":1,"fsid":"a3ab6672-193a-11f1-b81f-a119763c7190","modified":"2026-03-06T08:59:10.955430Z","created":"2026-03-06T08:59:10.955430Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-06T10:00:43.708 INFO:teuthology.orchestra.run.vm07.stderr:dumped monmap epoch 1 2026-03-06T10:00:43.879 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:43 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/3038824714' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T10:00:44.897 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-06T10:00:44.897 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph mon dump -f json 2026-03-06T10:00:45.216 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:45.250 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:45.625 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-06T10:00:45.625 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":1,"fsid":"a3ab6672-193a-11f1-b81f-a119763c7190","modified":"2026-03-06T08:59:10.955430Z","created":"2026-03-06T08:59:10.955430Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-06T10:00:45.625 INFO:teuthology.orchestra.run.vm07.stderr:dumped monmap epoch 1 2026-03-06T10:00:46.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:45 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/3146612032' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T10:00:46.809 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-06T10:00:46.810 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph mon dump -f json 2026-03-06T10:00:47.128 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:47.164 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:47.544 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-06T10:00:47.544 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":1,"fsid":"a3ab6672-193a-11f1-b81f-a119763c7190","modified":"2026-03-06T08:59:10.955430Z","created":"2026-03-06T08:59:10.955430Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-06T10:00:47.544 INFO:teuthology.orchestra.run.vm07.stderr:dumped monmap epoch 1 2026-03-06T10:00:47.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:47 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/3481333205' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T10:00:48.730 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-06T10:00:48.731 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph mon dump -f json 2026-03-06T10:00:49.012 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:48 vm01 ceph-mon[50931]: Active manager daemon vm01.ifwqbh restarted 2026-03-06T10:00:49.012 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:48 vm01 ceph-mon[50931]: Activating manager daemon vm01.ifwqbh 2026-03-06T10:00:49.012 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:48 vm01 ceph-mon[50931]: osdmap e5: 0 total, 0 up, 0 in 2026-03-06T10:00:49.012 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:48 vm01 ceph-mon[50931]: mgrmap e15: vm01.ifwqbh(active, starting, since 0.0062036s) 2026-03-06T10:00:49.012 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:48 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-06T10:00:49.012 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:48 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mgr metadata", "who": "vm01.ifwqbh", "id": "vm01.ifwqbh"}]: dispatch 2026-03-06T10:00:49.012 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:48 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-06T10:00:49.012 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:48 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-06T10:00:49.012 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:48 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-06T10:00:49.012 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:48 vm01 ceph-mon[50931]: Manager daemon vm01.ifwqbh is now available 2026-03-06T10:00:49.012 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:48 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:49.058 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:49.106 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:49.554 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-06T10:00:49.554 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":1,"fsid":"a3ab6672-193a-11f1-b81f-a119763c7190","modified":"2026-03-06T08:59:10.955430Z","created":"2026-03-06T08:59:10.955430Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-06T10:00:49.554 INFO:teuthology.orchestra.run.vm07.stderr:dumped monmap epoch 1 2026-03-06T10:00:50.148 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:49 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.ifwqbh/mirror_snapshot_schedule"}]: dispatch 2026-03-06T10:00:50.148 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:49 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T10:00:50.148 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:49 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T10:00:50.148 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:49 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm01.ifwqbh/trash_purge_schedule"}]: dispatch 2026-03-06T10:00:50.148 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:49 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/1002314578' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T10:00:50.148 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:49 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:50.148 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:49 vm01 ceph-mon[50931]: mgrmap e16: vm01.ifwqbh(active, since 1.00792s) 2026-03-06T10:00:50.723 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-06T10:00:50.723 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph mon dump -f json 2026-03-06T10:00:51.109 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:51.161 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:51.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:50 vm01 ceph-mon[50931]: [06/Mar/2026:09:00:49] ENGINE Bus STARTING 2026-03-06T10:00:51.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:50 vm01 ceph-mon[50931]: [06/Mar/2026:09:00:49] ENGINE Serving on https://192.168.123.101:7150 2026-03-06T10:00:51.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:50 vm01 ceph-mon[50931]: [06/Mar/2026:09:00:49] ENGINE Client ('192.168.123.101', 37686) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-06T10:00:51.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:50 vm01 ceph-mon[50931]: [06/Mar/2026:09:00:49] ENGINE Serving on http://192.168.123.101:8765 2026-03-06T10:00:51.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:50 vm01 ceph-mon[50931]: [06/Mar/2026:09:00:49] ENGINE Bus STARTED 2026-03-06T10:00:51.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:50 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:51.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:50 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:51.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:50 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:51.583 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-06T10:00:51.583 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":1,"fsid":"a3ab6672-193a-11f1-b81f-a119763c7190","modified":"2026-03-06T08:59:10.955430Z","created":"2026-03-06T08:59:10.955430Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-06T10:00:51.583 INFO:teuthology.orchestra.run.vm07.stderr:dumped monmap epoch 1 2026-03-06T10:00:52.741 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-06T10:00:52.741 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph mon dump -f json 2026-03-06T10:00:52.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:52 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:52.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:52 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:52.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:52 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-06T10:00:52.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:52 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:52.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:52 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/3323088564' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T10:00:52.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:52 vm01 ceph-mon[50931]: mgrmap e17: vm01.ifwqbh(active, since 2s) 2026-03-06T10:00:53.098 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:53.152 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T10:00:53.576 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-06T10:00:53.576 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":1,"fsid":"a3ab6672-193a-11f1-b81f-a119763c7190","modified":"2026-03-06T08:59:10.955430Z","created":"2026-03-06T08:59:10.955430Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-06T10:00:53.576 INFO:teuthology.orchestra.run.vm07.stderr:dumped monmap epoch 1 2026-03-06T10:00:53.770 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:53 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/3259534482' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T10:00:53.770 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:53 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:53.770 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:53 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:53.770 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:53 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:53.770 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:53 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:53.770 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:53 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config rm", "who": "osd/host:vm07", "name": "osd_memory_target"}]: dispatch 2026-03-06T10:00:53.770 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:53 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:00:53.770 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:53 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T10:00:54.746 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-06T10:00:54.747 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph mon dump -f json 2026-03-06T10:00:54.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:54 vm01 ceph-mon[50931]: Updating vm01:/etc/ceph/ceph.conf 2026-03-06T10:00:54.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:54 vm01 ceph-mon[50931]: Updating vm07:/etc/ceph/ceph.conf 2026-03-06T10:00:54.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:54 vm01 ceph-mon[50931]: Updating vm01:/var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/config/ceph.conf 2026-03-06T10:00:54.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:54 vm01 ceph-mon[50931]: Updating vm07:/var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/config/ceph.conf 2026-03-06T10:00:54.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:54 vm01 ceph-mon[50931]: Updating vm01:/etc/ceph/ceph.client.admin.keyring 2026-03-06T10:00:54.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:54 vm01 ceph-mon[50931]: Updating vm07:/etc/ceph/ceph.client.admin.keyring 2026-03-06T10:00:54.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:54 vm01 ceph-mon[50931]: Updating vm01:/var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/config/ceph.client.admin.keyring 2026-03-06T10:00:54.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:54 vm01 ceph-mon[50931]: Updating vm07:/var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/config/ceph.client.admin.keyring 2026-03-06T10:00:54.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:54 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:54.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:54 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:54.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:54 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:54.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:54 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:54.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:54 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:54.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:54 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm07", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-06T10:00:54.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:54 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd='[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm07", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]': finished 2026-03-06T10:00:54.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:54 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:00:55.275 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/config/ceph.conf 2026-03-06T10:00:55.941 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-06T10:00:55.941 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":1,"fsid":"a3ab6672-193a-11f1-b81f-a119763c7190","modified":"2026-03-06T08:59:10.955430Z","created":"2026-03-06T08:59:10.955430Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-06T10:00:55.941 INFO:teuthology.orchestra.run.vm07.stderr:dumped monmap epoch 1 2026-03-06T10:00:56.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:55 vm01 ceph-mon[50931]: Deploying daemon ceph-exporter.vm07 on vm07 2026-03-06T10:00:57.124 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-06T10:00:57.125 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph mon dump -f json 2026-03-06T10:00:57.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:56 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:57.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:56 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:57.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:56 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:57.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:56 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:57.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:56 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm07", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-06T10:00:57.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:56 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd='[{"prefix": "auth get-or-create", "entity": "client.crash.vm07", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]': finished 2026-03-06T10:00:57.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:56 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:00:57.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:56 vm01 ceph-mon[50931]: Deploying daemon crash.vm07 on vm07 2026-03-06T10:00:57.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:56 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/1955034862' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T10:00:57.530 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/config/ceph.conf 2026-03-06T10:00:58.039 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-06T10:00:58.039 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":1,"fsid":"a3ab6672-193a-11f1-b81f-a119763c7190","modified":"2026-03-06T08:59:10.955430Z","created":"2026-03-06T08:59:10.955430Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-06T10:00:58.039 INFO:teuthology.orchestra.run.vm07.stderr:dumped monmap epoch 1 2026-03-06T10:00:58.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:57 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:58.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:57 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:58.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:57 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:58.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:57 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:58.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:57 vm01 ceph-mon[50931]: Deploying daemon node-exporter.vm07 on vm07 2026-03-06T10:00:59.197 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-06T10:00:59.197 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph mon dump -f json 2026-03-06T10:00:59.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:58 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/1371810333' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T10:00:59.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:00:58 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:00:59.541 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/config/ceph.conf 2026-03-06T10:00:59.988 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-06T10:00:59.988 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":1,"fsid":"a3ab6672-193a-11f1-b81f-a119763c7190","modified":"2026-03-06T08:59:10.955430Z","created":"2026-03-06T08:59:10.955430Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-06T10:00:59.989 INFO:teuthology.orchestra.run.vm07.stderr:dumped monmap epoch 1 2026-03-06T10:01:00.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:00 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/2424712175' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T10:01:01.246 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-06T10:01:01.246 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph mon dump -f json 2026-03-06T10:01:01.748 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/config/ceph.conf 2026-03-06T10:01:01.961 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:01 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:01.961 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:01 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:01.961 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:01 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:01.961 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:01 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:01.961 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:01 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm07.myglqt", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T10:01:01.961 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:01 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.vm07.myglqt", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-06T10:01:01.961 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:01 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T10:01:01.962 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:01 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:01.962 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:01 vm01 ceph-mon[50931]: Deploying daemon mgr.vm07.myglqt on vm07 2026-03-06T10:01:02.416 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-06T10:01:02.416 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":1,"fsid":"a3ab6672-193a-11f1-b81f-a119763c7190","modified":"2026-03-06T08:59:10.955430Z","created":"2026-03-06T08:59:10.955430Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-06T10:01:02.416 INFO:teuthology.orchestra.run.vm07.stderr:dumped monmap epoch 1 2026-03-06T10:01:03.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:02 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:03.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:02 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:03.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:02 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:03.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:02 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:03.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:02 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T10:01:03.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:02 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:03.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:02 vm01 ceph-mon[50931]: Deploying daemon mon.vm07 on vm07 2026-03-06T10:01:03.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:02 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/2439166692' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T10:01:03.591 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-06T10:01:03.591 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph mon dump -f json 2026-03-06T10:01:03.992 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm07/config 2026-03-06T10:01:08.947 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-06T10:01:08.948 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":2,"fsid":"a3ab6672-193a-11f1-b81f-a119763c7190","modified":"2026-03-06T09:01:03.514609Z","created":"2026-03-06T08:59:10.955430Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm01","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"vm07","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:3300","nonce":0},{"type":"v1","addr":"192.168.123.107:6789","nonce":0}]},"addr":"192.168.123.107:6789/0","public_addr":"192.168.123.107:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1]} 2026-03-06T10:01:08.948 INFO:teuthology.orchestra.run.vm07.stderr:dumped monmap epoch 2 2026-03-06T10:01:09.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:08 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-06T10:01:09.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:08 vm01 ceph-mon[50931]: mon.vm01 calling monitor election 2026-03-06T10:01:09.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:08 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-06T10:01:09.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:08 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T10:01:09.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:08 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T10:01:09.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:08 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-06T10:01:09.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:08 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-06T10:01:09.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:08 vm01 ceph-mon[50931]: mon.vm07 calling monitor election 2026-03-06T10:01:09.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:08 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-06T10:01:09.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:08 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-06T10:01:09.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:08 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-06T10:01:09.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:08 vm01 ceph-mon[50931]: mon.vm01 is new leader, mons vm01,vm07 in quorum (ranks 0,1) 2026-03-06T10:01:09.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:08 vm01 ceph-mon[50931]: monmap epoch 2 2026-03-06T10:01:09.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:08 vm01 ceph-mon[50931]: fsid a3ab6672-193a-11f1-b81f-a119763c7190 2026-03-06T10:01:09.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:08 vm01 ceph-mon[50931]: last_changed 2026-03-06T09:01:03.514609+0000 2026-03-06T10:01:09.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:08 vm01 ceph-mon[50931]: created 2026-03-06T08:59:10.955430+0000 2026-03-06T10:01:09.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:08 vm01 ceph-mon[50931]: min_mon_release 19 (squid) 2026-03-06T10:01:09.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:08 vm01 ceph-mon[50931]: election_strategy: 1 2026-03-06T10:01:09.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:08 vm01 ceph-mon[50931]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.vm01 2026-03-06T10:01:09.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:08 vm01 ceph-mon[50931]: 1: [v2:192.168.123.107:3300/0,v1:192.168.123.107:6789/0] mon.vm07 2026-03-06T10:01:09.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:08 vm01 ceph-mon[50931]: fsmap 2026-03-06T10:01:09.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:08 vm01 ceph-mon[50931]: osdmap e5: 0 total, 0 up, 0 in 2026-03-06T10:01:09.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:08 vm01 ceph-mon[50931]: mgrmap e17: vm01.ifwqbh(active, since 19s) 2026-03-06T10:01:09.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:08 vm01 ceph-mon[50931]: overall HEALTH_OK 2026-03-06T10:01:09.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:08 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:09.186 INFO:tasks.cephadm:Generating final ceph.conf file... 2026-03-06T10:01:09.186 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph config generate-minimal-conf 2026-03-06T10:01:09.541 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:01:09.854 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:09 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-06T10:01:09.924 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:09 vm01 ceph-mon[50931]: pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:09.924 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:09 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/174158741' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T10:01:09.924 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:09 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-06T10:01:09.924 INFO:teuthology.orchestra.run.vm01.stdout:# minimal ceph.conf for a3ab6672-193a-11f1-b81f-a119763c7190 2026-03-06T10:01:09.924 INFO:teuthology.orchestra.run.vm01.stdout:[global] 2026-03-06T10:01:09.924 INFO:teuthology.orchestra.run.vm01.stdout: fsid = a3ab6672-193a-11f1-b81f-a119763c7190 2026-03-06T10:01:09.924 INFO:teuthology.orchestra.run.vm01.stdout: mon_host = [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] [v2:192.168.123.107:3300/0,v1:192.168.123.107:6789/0] 2026-03-06T10:01:10.090 INFO:tasks.cephadm:Distributing (final) config and client.admin keyring... 2026-03-06T10:01:10.090 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T10:01:10.090 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/ceph/ceph.conf 2026-03-06T10:01:10.119 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T10:01:10.119 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-06T10:01:10.187 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-06T10:01:10.187 DEBUG:teuthology.orchestra.run.vm07:> sudo dd of=/etc/ceph/ceph.conf 2026-03-06T10:01:10.214 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-06T10:01:10.214 DEBUG:teuthology.orchestra.run.vm07:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-06T10:01:10.279 INFO:tasks.cephadm:Deploying OSDs... 2026-03-06T10:01:10.279 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T10:01:10.279 DEBUG:teuthology.orchestra.run.vm01:> dd if=/scratch_devs of=/dev/stdout 2026-03-06T10:01:10.296 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T10:01:10.296 DEBUG:teuthology.orchestra.run.vm01:> ls /dev/[sv]d? 2026-03-06T10:01:10.353 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vda 2026-03-06T10:01:10.353 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vdb 2026-03-06T10:01:10.353 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vdc 2026-03-06T10:01:10.353 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vdd 2026-03-06T10:01:10.353 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vde 2026-03-06T10:01:10.353 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-06T10:01:10.353 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-06T10:01:10.353 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vdb 2026-03-06T10:01:10.422 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vdb 2026-03-06T10:01:10.422 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T10:01:10.422 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 221 Links: 1 Device type: fc,10 2026-03-06T10:01:10.422 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T10:01:10.422 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T10:01:10.422 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-06 10:00:12.778389890 +0100 2026-03-06T10:01:10.422 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-06 09:58:25.674648503 +0100 2026-03-06T10:01:10.422 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-06 09:58:25.674648503 +0100 2026-03-06T10:01:10.422 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-06 09:55:38.226000000 +0100 2026-03-06T10:01:10.422 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-06T10:01:10.510 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-06T10:01:10.510 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-06T10:01:10.511 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.000118402 s, 4.3 MB/s 2026-03-06T10:01:10.516 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-06T10:01:10.543 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vdc 2026-03-06T10:01:10.602 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vdc 2026-03-06T10:01:10.602 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T10:01:10.602 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 224 Links: 1 Device type: fc,20 2026-03-06T10:01:10.602 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T10:01:10.602 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T10:01:10.602 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-06 10:00:12.837389788 +0100 2026-03-06T10:01:10.602 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-06 09:58:25.728648511 +0100 2026-03-06T10:01:10.602 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-06 09:58:25.728648511 +0100 2026-03-06T10:01:10.602 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-06 09:55:38.230000000 +0100 2026-03-06T10:01:10.603 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-06T10:01:10.656 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:10 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:10.656 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:10 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:10.656 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:10 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:10.656 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:10 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T10:01:10.656 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:10 vm07 ceph-mon[55141]: Updating vm01:/etc/ceph/ceph.conf 2026-03-06T10:01:10.656 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:10 vm07 ceph-mon[55141]: Updating vm07:/etc/ceph/ceph.conf 2026-03-06T10:01:10.656 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:10 vm07 ceph-mon[55141]: Updating vm07:/var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/config/ceph.conf 2026-03-06T10:01:10.656 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:10 vm07 ceph-mon[55141]: Updating vm01:/var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/config/ceph.conf 2026-03-06T10:01:10.656 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:10 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/719184830' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:10.657 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:10 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:10.657 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:10 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:10.657 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:10 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:10.657 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:10 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:10.657 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:10 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:10.657 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:10 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T10:01:10.657 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:10 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T10:01:10.657 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:10 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:10.687 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-06T10:01:10.687 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-06T10:01:10.687 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.000103614 s, 4.9 MB/s 2026-03-06T10:01:10.688 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-06T10:01:10.709 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vdd 2026-03-06T10:01:10.766 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vdd 2026-03-06T10:01:10.766 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T10:01:10.767 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 225 Links: 1 Device type: fc,30 2026-03-06T10:01:10.767 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T10:01:10.767 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T10:01:10.767 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-06 10:00:12.910389662 +0100 2026-03-06T10:01:10.767 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-06 09:58:25.691648506 +0100 2026-03-06T10:01:10.767 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-06 09:58:25.691648506 +0100 2026-03-06T10:01:10.767 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-06 09:55:38.241000000 +0100 2026-03-06T10:01:10.767 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-06T10:01:10.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:10 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:10.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:10 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:10.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:10 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:10.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:10 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T10:01:10.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:10 vm01 ceph-mon[50931]: Updating vm01:/etc/ceph/ceph.conf 2026-03-06T10:01:10.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:10 vm01 ceph-mon[50931]: Updating vm07:/etc/ceph/ceph.conf 2026-03-06T10:01:10.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:10 vm01 ceph-mon[50931]: Updating vm07:/var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/config/ceph.conf 2026-03-06T10:01:10.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:10 vm01 ceph-mon[50931]: Updating vm01:/var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/config/ceph.conf 2026-03-06T10:01:10.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:10 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/719184830' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:10.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:10 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:10.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:10 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:10.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:10 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:10.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:10 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:10.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:10 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:10.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:10 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T10:01:10.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:10 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T10:01:10.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:10 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:10.794 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-06T10:01:10.794 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-06T10:01:10.794 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.000124232 s, 4.1 MB/s 2026-03-06T10:01:10.795 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-06T10:01:10.856 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vde 2026-03-06T10:01:10.914 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vde 2026-03-06T10:01:10.914 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T10:01:10.914 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 247 Links: 1 Device type: fc,40 2026-03-06T10:01:10.914 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T10:01:10.914 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T10:01:10.914 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-06 10:00:12.956389583 +0100 2026-03-06T10:01:10.914 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-06 09:58:25.719648510 +0100 2026-03-06T10:01:10.914 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-06 09:58:25.719648510 +0100 2026-03-06T10:01:10.914 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-06 09:55:38.278000000 +0100 2026-03-06T10:01:10.914 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-06T10:01:10.981 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-06T10:01:10.981 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-06T10:01:10.981 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.000211116 s, 2.4 MB/s 2026-03-06T10:01:10.982 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-06T10:01:11.046 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-06T10:01:11.046 DEBUG:teuthology.orchestra.run.vm07:> dd if=/scratch_devs of=/dev/stdout 2026-03-06T10:01:11.072 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T10:01:11.072 DEBUG:teuthology.orchestra.run.vm07:> ls /dev/[sv]d? 2026-03-06T10:01:11.129 INFO:teuthology.orchestra.run.vm07.stdout:/dev/vda 2026-03-06T10:01:11.129 INFO:teuthology.orchestra.run.vm07.stdout:/dev/vdb 2026-03-06T10:01:11.130 INFO:teuthology.orchestra.run.vm07.stdout:/dev/vdc 2026-03-06T10:01:11.130 INFO:teuthology.orchestra.run.vm07.stdout:/dev/vdd 2026-03-06T10:01:11.130 INFO:teuthology.orchestra.run.vm07.stdout:/dev/vde 2026-03-06T10:01:11.130 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-06T10:01:11.130 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-06T10:01:11.130 DEBUG:teuthology.orchestra.run.vm07:> stat /dev/vdb 2026-03-06T10:01:11.188 INFO:teuthology.orchestra.run.vm07.stdout: File: /dev/vdb 2026-03-06T10:01:11.191 INFO:teuthology.orchestra.run.vm07.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T10:01:11.191 INFO:teuthology.orchestra.run.vm07.stdout:Device: 6h/6d Inode: 223 Links: 1 Device type: fc,10 2026-03-06T10:01:11.191 INFO:teuthology.orchestra.run.vm07.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T10:01:11.191 INFO:teuthology.orchestra.run.vm07.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T10:01:11.191 INFO:teuthology.orchestra.run.vm07.stdout:Access: 2026-03-06 10:00:53.285889076 +0100 2026-03-06T10:01:11.191 INFO:teuthology.orchestra.run.vm07.stdout:Modify: 2026-03-06 09:58:26.184695948 +0100 2026-03-06T10:01:11.191 INFO:teuthology.orchestra.run.vm07.stdout:Change: 2026-03-06 09:58:26.184695948 +0100 2026-03-06T10:01:11.191 INFO:teuthology.orchestra.run.vm07.stdout: Birth: 2026-03-06 09:55:13.244000000 +0100 2026-03-06T10:01:11.191 DEBUG:teuthology.orchestra.run.vm07:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-06T10:01:11.255 INFO:teuthology.orchestra.run.vm07.stderr:1+0 records in 2026-03-06T10:01:11.255 INFO:teuthology.orchestra.run.vm07.stderr:1+0 records out 2026-03-06T10:01:11.255 INFO:teuthology.orchestra.run.vm07.stderr:512 bytes copied, 0.000137258 s, 3.7 MB/s 2026-03-06T10:01:11.257 DEBUG:teuthology.orchestra.run.vm07:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-06T10:01:11.315 DEBUG:teuthology.orchestra.run.vm07:> stat /dev/vdc 2026-03-06T10:01:11.374 INFO:teuthology.orchestra.run.vm07.stdout: File: /dev/vdc 2026-03-06T10:01:11.374 INFO:teuthology.orchestra.run.vm07.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T10:01:11.374 INFO:teuthology.orchestra.run.vm07.stdout:Device: 6h/6d Inode: 246 Links: 1 Device type: fc,20 2026-03-06T10:01:11.374 INFO:teuthology.orchestra.run.vm07.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T10:01:11.374 INFO:teuthology.orchestra.run.vm07.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T10:01:11.374 INFO:teuthology.orchestra.run.vm07.stdout:Access: 2026-03-06 10:00:53.316889130 +0100 2026-03-06T10:01:11.374 INFO:teuthology.orchestra.run.vm07.stdout:Modify: 2026-03-06 09:58:26.179695947 +0100 2026-03-06T10:01:11.374 INFO:teuthology.orchestra.run.vm07.stdout:Change: 2026-03-06 09:58:26.179695947 +0100 2026-03-06T10:01:11.374 INFO:teuthology.orchestra.run.vm07.stdout: Birth: 2026-03-06 09:55:13.266000000 +0100 2026-03-06T10:01:11.374 DEBUG:teuthology.orchestra.run.vm07:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-06T10:01:11.442 INFO:teuthology.orchestra.run.vm07.stderr:1+0 records in 2026-03-06T10:01:11.442 INFO:teuthology.orchestra.run.vm07.stderr:1+0 records out 2026-03-06T10:01:11.442 INFO:teuthology.orchestra.run.vm07.stderr:512 bytes copied, 0.000104335 s, 4.9 MB/s 2026-03-06T10:01:11.443 DEBUG:teuthology.orchestra.run.vm07:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-06T10:01:11.506 DEBUG:teuthology.orchestra.run.vm07:> stat /dev/vdd 2026-03-06T10:01:11.564 INFO:teuthology.orchestra.run.vm07.stdout: File: /dev/vdd 2026-03-06T10:01:11.564 INFO:teuthology.orchestra.run.vm07.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T10:01:11.564 INFO:teuthology.orchestra.run.vm07.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-06T10:01:11.564 INFO:teuthology.orchestra.run.vm07.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T10:01:11.564 INFO:teuthology.orchestra.run.vm07.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T10:01:11.565 INFO:teuthology.orchestra.run.vm07.stdout:Access: 2026-03-06 10:00:53.347889184 +0100 2026-03-06T10:01:11.565 INFO:teuthology.orchestra.run.vm07.stdout:Modify: 2026-03-06 09:58:26.185695948 +0100 2026-03-06T10:01:11.565 INFO:teuthology.orchestra.run.vm07.stdout:Change: 2026-03-06 09:58:26.185695948 +0100 2026-03-06T10:01:11.565 INFO:teuthology.orchestra.run.vm07.stdout: Birth: 2026-03-06 09:55:13.281000000 +0100 2026-03-06T10:01:11.565 DEBUG:teuthology.orchestra.run.vm07:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-06T10:01:11.629 INFO:teuthology.orchestra.run.vm07.stderr:1+0 records in 2026-03-06T10:01:11.629 INFO:teuthology.orchestra.run.vm07.stderr:1+0 records out 2026-03-06T10:01:11.629 INFO:teuthology.orchestra.run.vm07.stderr:512 bytes copied, 0.000135122 s, 3.8 MB/s 2026-03-06T10:01:11.630 DEBUG:teuthology.orchestra.run.vm07:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-06T10:01:11.698 DEBUG:teuthology.orchestra.run.vm07:> stat /dev/vde 2026-03-06T10:01:11.757 INFO:teuthology.orchestra.run.vm07.stdout: File: /dev/vde 2026-03-06T10:01:11.758 INFO:teuthology.orchestra.run.vm07.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T10:01:11.758 INFO:teuthology.orchestra.run.vm07.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-06T10:01:11.758 INFO:teuthology.orchestra.run.vm07.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T10:01:11.758 INFO:teuthology.orchestra.run.vm07.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T10:01:11.758 INFO:teuthology.orchestra.run.vm07.stdout:Access: 2026-03-06 10:00:53.380889242 +0100 2026-03-06T10:01:11.758 INFO:teuthology.orchestra.run.vm07.stdout:Modify: 2026-03-06 09:58:26.219695952 +0100 2026-03-06T10:01:11.758 INFO:teuthology.orchestra.run.vm07.stdout:Change: 2026-03-06 09:58:26.219695952 +0100 2026-03-06T10:01:11.758 INFO:teuthology.orchestra.run.vm07.stdout: Birth: 2026-03-06 09:55:13.322000000 +0100 2026-03-06T10:01:11.758 DEBUG:teuthology.orchestra.run.vm07:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-06T10:01:11.826 INFO:teuthology.orchestra.run.vm07.stderr:1+0 records in 2026-03-06T10:01:11.826 INFO:teuthology.orchestra.run.vm07.stderr:1+0 records out 2026-03-06T10:01:11.826 INFO:teuthology.orchestra.run.vm07.stderr:512 bytes copied, 0.000164357 s, 3.1 MB/s 2026-03-06T10:01:11.828 DEBUG:teuthology.orchestra.run.vm07:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-06T10:01:11.887 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph orch apply osd --all-available-devices 2026-03-06T10:01:11.947 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:11 vm07 ceph-mon[55141]: Reconfiguring mon.vm01 (unknown last config time)... 2026-03-06T10:01:11.947 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:11 vm07 ceph-mon[55141]: Reconfiguring daemon mon.vm01 on vm01 2026-03-06T10:01:11.947 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:11 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:11.947 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:11 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:11.947 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:11 vm07 ceph-mon[55141]: Reconfiguring mgr.vm01.ifwqbh (unknown last config time)... 2026-03-06T10:01:11.947 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:11 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm01.ifwqbh", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T10:01:11.947 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:11 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T10:01:11.947 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:11 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:11.947 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:11 vm07 ceph-mon[55141]: Reconfiguring daemon mgr.vm01.ifwqbh on vm01 2026-03-06T10:01:11.947 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:11 vm07 ceph-mon[55141]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:11.947 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:11 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:11.947 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:11 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:11.947 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:11 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm01", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-06T10:01:11.947 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:11 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:11.951 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:11 vm01 ceph-mon[50931]: Reconfiguring mon.vm01 (unknown last config time)... 2026-03-06T10:01:11.951 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:11 vm01 ceph-mon[50931]: Reconfiguring daemon mon.vm01 on vm01 2026-03-06T10:01:11.951 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:11 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:11.951 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:11 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:11.951 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:11 vm01 ceph-mon[50931]: Reconfiguring mgr.vm01.ifwqbh (unknown last config time)... 2026-03-06T10:01:11.951 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:11 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm01.ifwqbh", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T10:01:11.951 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:11 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T10:01:11.951 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:11 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:11.951 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:11 vm01 ceph-mon[50931]: Reconfiguring daemon mgr.vm01.ifwqbh on vm01 2026-03-06T10:01:11.951 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:11 vm01 ceph-mon[50931]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:11.951 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:11 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:11.951 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:11 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:11.951 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:11 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm01", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-06T10:01:11.951 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:11 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:12.282 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm07/config 2026-03-06T10:01:12.654 INFO:teuthology.orchestra.run.vm07.stdout:Scheduled osd.all-available-devices update... 2026-03-06T10:01:12.835 INFO:tasks.cephadm:Waiting for 8 OSDs to come up... 2026-03-06T10:01:12.835 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd stat -f json 2026-03-06T10:01:13.155 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:12 vm01 ceph-mon[50931]: Reconfiguring ceph-exporter.vm01 (monmap changed)... 2026-03-06T10:01:13.155 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:12 vm01 ceph-mon[50931]: Reconfiguring daemon ceph-exporter.vm01 on vm01 2026-03-06T10:01:13.155 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:12 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:13.155 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:12 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:13.155 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:12 vm01 ceph-mon[50931]: Reconfiguring crash.vm01 (monmap changed)... 2026-03-06T10:01:13.155 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:12 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm01", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-06T10:01:13.155 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:12 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:13.155 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:12 vm01 ceph-mon[50931]: Reconfiguring daemon crash.vm01 on vm01 2026-03-06T10:01:13.155 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:12 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:13.155 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:12 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:13.155 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:12 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:13.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:12 vm07 ceph-mon[55141]: Reconfiguring ceph-exporter.vm01 (monmap changed)... 2026-03-06T10:01:13.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:12 vm07 ceph-mon[55141]: Reconfiguring daemon ceph-exporter.vm01 on vm01 2026-03-06T10:01:13.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:12 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:13.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:12 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:13.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:12 vm07 ceph-mon[55141]: Reconfiguring crash.vm01 (monmap changed)... 2026-03-06T10:01:13.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:12 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm01", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-06T10:01:13.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:12 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:13.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:12 vm07 ceph-mon[55141]: Reconfiguring daemon crash.vm01 on vm01 2026-03-06T10:01:13.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:12 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:13.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:12 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:13.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:12 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:13.270 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:01:13.641 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:01:13.820 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":5,"num_osds":0,"num_up_osds":0,"osd_up_since":0,"num_in_osds":0,"osd_in_since":0,"num_remapped_pgs":0} 2026-03-06T10:01:14.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:14 vm01 ceph-mon[50931]: Reconfiguring alertmanager.vm01 (dependencies changed)... 2026-03-06T10:01:14.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:14 vm01 ceph-mon[50931]: Reconfiguring daemon alertmanager.vm01 on vm01 2026-03-06T10:01:14.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:14 vm01 ceph-mon[50931]: from='client.14264 -' entity='client.admin' cmd=[{"prefix": "orch apply osd", "all_available_devices": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:01:14.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:14 vm01 ceph-mon[50931]: Marking host: vm01 for OSDSpec preview refresh. 2026-03-06T10:01:14.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:14 vm01 ceph-mon[50931]: Marking host: vm07 for OSDSpec preview refresh. 2026-03-06T10:01:14.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:14 vm01 ceph-mon[50931]: Saving service osd.all-available-devices spec with placement * 2026-03-06T10:01:14.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:14 vm01 ceph-mon[50931]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:14.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:14 vm01 ceph-mon[50931]: Standby manager daemon vm07.myglqt started 2026-03-06T10:01:14.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:14 vm01 ceph-mon[50931]: from='mgr.? 192.168.123.107:0/10923400' entity='mgr.vm07.myglqt' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm07.myglqt/crt"}]: dispatch 2026-03-06T10:01:14.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:14 vm01 ceph-mon[50931]: from='mgr.? 192.168.123.107:0/10923400' entity='mgr.vm07.myglqt' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-06T10:01:14.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:14 vm01 ceph-mon[50931]: from='mgr.? 192.168.123.107:0/10923400' entity='mgr.vm07.myglqt' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm07.myglqt/key"}]: dispatch 2026-03-06T10:01:14.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:14 vm01 ceph-mon[50931]: from='mgr.? 192.168.123.107:0/10923400' entity='mgr.vm07.myglqt' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-06T10:01:14.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:14 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/2264713480' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:14.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:14 vm07 ceph-mon[55141]: Reconfiguring alertmanager.vm01 (dependencies changed)... 2026-03-06T10:01:14.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:14 vm07 ceph-mon[55141]: Reconfiguring daemon alertmanager.vm01 on vm01 2026-03-06T10:01:14.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:14 vm07 ceph-mon[55141]: from='client.14264 -' entity='client.admin' cmd=[{"prefix": "orch apply osd", "all_available_devices": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:01:14.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:14 vm07 ceph-mon[55141]: Marking host: vm01 for OSDSpec preview refresh. 2026-03-06T10:01:14.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:14 vm07 ceph-mon[55141]: Marking host: vm07 for OSDSpec preview refresh. 2026-03-06T10:01:14.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:14 vm07 ceph-mon[55141]: Saving service osd.all-available-devices spec with placement * 2026-03-06T10:01:14.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:14 vm07 ceph-mon[55141]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:14.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:14 vm07 ceph-mon[55141]: Standby manager daemon vm07.myglqt started 2026-03-06T10:01:14.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:14 vm07 ceph-mon[55141]: from='mgr.? 192.168.123.107:0/10923400' entity='mgr.vm07.myglqt' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm07.myglqt/crt"}]: dispatch 2026-03-06T10:01:14.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:14 vm07 ceph-mon[55141]: from='mgr.? 192.168.123.107:0/10923400' entity='mgr.vm07.myglqt' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-06T10:01:14.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:14 vm07 ceph-mon[55141]: from='mgr.? 192.168.123.107:0/10923400' entity='mgr.vm07.myglqt' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm07.myglqt/key"}]: dispatch 2026-03-06T10:01:14.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:14 vm07 ceph-mon[55141]: from='mgr.? 192.168.123.107:0/10923400' entity='mgr.vm07.myglqt' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-06T10:01:14.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:14 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/2264713480' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:14.821 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd stat -f json 2026-03-06T10:01:15.197 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:01:15.452 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:15 vm01 ceph-mon[50931]: mgrmap e18: vm01.ifwqbh(active, since 25s), standbys: vm07.myglqt 2026-03-06T10:01:15.452 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:15 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mgr metadata", "who": "vm07.myglqt", "id": "vm07.myglqt"}]: dispatch 2026-03-06T10:01:15.453 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:15 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:15.453 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:15 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:15.642 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:01:15.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:15 vm07 ceph-mon[55141]: mgrmap e18: vm01.ifwqbh(active, since 25s), standbys: vm07.myglqt 2026-03-06T10:01:15.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:15 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mgr metadata", "who": "vm07.myglqt", "id": "vm07.myglqt"}]: dispatch 2026-03-06T10:01:15.697 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:15 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:15.697 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:15 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:15.991 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":5,"num_osds":0,"num_up_osds":0,"osd_up_since":0,"num_in_osds":0,"osd_in_since":0,"num_remapped_pgs":0} 2026-03-06T10:01:16.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:16 vm01 ceph-mon[50931]: Reconfiguring grafana.vm01 (dependencies changed)... 2026-03-06T10:01:16.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:16 vm01 ceph-mon[50931]: Reconfiguring daemon grafana.vm01 on vm01 2026-03-06T10:01:16.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:16 vm01 ceph-mon[50931]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:16.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:16 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/1671464016' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:16.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:16 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:16.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:16 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:16.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:16 vm01 ceph-mon[50931]: Reconfiguring prometheus.vm01 (dependencies changed)... 2026-03-06T10:01:16.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:16 vm01 ceph-mon[50931]: Reconfiguring daemon prometheus.vm01 on vm01 2026-03-06T10:01:16.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:16 vm07 ceph-mon[55141]: Reconfiguring grafana.vm01 (dependencies changed)... 2026-03-06T10:01:16.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:16 vm07 ceph-mon[55141]: Reconfiguring daemon grafana.vm01 on vm01 2026-03-06T10:01:16.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:16 vm07 ceph-mon[55141]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:16.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:16 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/1671464016' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:16.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:16 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:16.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:16 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:16.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:16 vm07 ceph-mon[55141]: Reconfiguring prometheus.vm01 (dependencies changed)... 2026-03-06T10:01:16.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:16 vm07 ceph-mon[55141]: Reconfiguring daemon prometheus.vm01 on vm01 2026-03-06T10:01:16.992 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd stat -f json 2026-03-06T10:01:17.407 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:01:17.489 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:17 vm01 ceph-mon[50931]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:17.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:17 vm07 ceph-mon[55141]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:17.756 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:01:17.939 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":5,"num_osds":0,"num_up_osds":0,"osd_up_since":0,"num_in_osds":0,"osd_in_since":0,"num_remapped_pgs":0} 2026-03-06T10:01:18.697 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:18 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:18.697 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:18 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:18.697 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:18 vm07 ceph-mon[55141]: Reconfiguring ceph-exporter.vm07 (monmap changed)... 2026-03-06T10:01:18.697 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:18 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm07", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-06T10:01:18.698 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:18 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:18.698 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:18 vm07 ceph-mon[55141]: Reconfiguring daemon ceph-exporter.vm07 on vm07 2026-03-06T10:01:18.698 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:18 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/2229039120' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:18.698 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:18 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:18.698 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:18 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:18.698 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:18 vm07 ceph-mon[55141]: Reconfiguring crash.vm07 (monmap changed)... 2026-03-06T10:01:18.698 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:18 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm07", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-06T10:01:18.698 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:18 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:18.698 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:18 vm07 ceph-mon[55141]: Reconfiguring daemon crash.vm07 on vm07 2026-03-06T10:01:18.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:18 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:18.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:18 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:18.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:18 vm01 ceph-mon[50931]: Reconfiguring ceph-exporter.vm07 (monmap changed)... 2026-03-06T10:01:18.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:18 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm07", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-06T10:01:18.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:18 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:18.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:18 vm01 ceph-mon[50931]: Reconfiguring daemon ceph-exporter.vm07 on vm07 2026-03-06T10:01:18.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:18 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/2229039120' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:18.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:18 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:18.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:18 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:18.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:18 vm01 ceph-mon[50931]: Reconfiguring crash.vm07 (monmap changed)... 2026-03-06T10:01:18.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:18 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm07", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-06T10:01:18.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:18 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:18.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:18 vm01 ceph-mon[50931]: Reconfiguring daemon crash.vm07 on vm07 2026-03-06T10:01:18.939 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd stat -f json 2026-03-06T10:01:19.283 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:01:19.625 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:01:19.783 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":5,"num_osds":0,"num_up_osds":0,"osd_up_since":0,"num_in_osds":0,"osd_in_since":0,"num_remapped_pgs":0} 2026-03-06T10:01:19.894 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:19 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:19.894 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:19 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:19.894 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:19 vm01 ceph-mon[50931]: Reconfiguring mgr.vm07.myglqt (monmap changed)... 2026-03-06T10:01:19.894 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:19 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm07.myglqt", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T10:01:19.894 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:19 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T10:01:19.894 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:19 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:19.894 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:19 vm01 ceph-mon[50931]: Reconfiguring daemon mgr.vm07.myglqt on vm07 2026-03-06T10:01:19.894 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:19 vm01 ceph-mon[50931]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:19.895 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:19 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T10:01:19.895 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:19 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:19.895 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:19 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:19.895 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:19 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T10:01:19.895 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:19 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T10:01:19.895 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:19 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:19.946 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:19 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:19.946 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:19 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:19.946 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:19 vm07 ceph-mon[55141]: Reconfiguring mgr.vm07.myglqt (monmap changed)... 2026-03-06T10:01:19.946 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:19 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm07.myglqt", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T10:01:19.946 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:19 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T10:01:19.946 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:19 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:19.946 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:19 vm07 ceph-mon[55141]: Reconfiguring daemon mgr.vm07.myglqt on vm07 2026-03-06T10:01:19.946 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:19 vm07 ceph-mon[55141]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:19.946 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:19 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T10:01:19.946 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:19 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:19.946 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:19 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:19.946 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:19 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T10:01:19.946 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:19 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T10:01:19.946 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:19 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:20.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:20 vm01 ceph-mon[50931]: Reconfiguring mon.vm07 (monmap changed)... 2026-03-06T10:01:20.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:20 vm01 ceph-mon[50931]: Reconfiguring daemon mon.vm07 on vm07 2026-03-06T10:01:20.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:20 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/3538377112' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:20.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:20 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:20.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:20 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:20.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:20 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-06T10:01:20.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:20 vm01 ceph-mon[50931]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-06T10:01:20.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:20 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm01.local:9093"}]: dispatch 2026-03-06T10:01:20.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:20 vm01 ceph-mon[50931]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm01.local:9093"}]: dispatch 2026-03-06T10:01:20.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:20 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:20.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:20 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-06T10:01:20.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:20 vm01 ceph-mon[50931]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-06T10:01:20.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:20 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm01.local:3000"}]: dispatch 2026-03-06T10:01:20.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:20 vm01 ceph-mon[50931]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm01.local:3000"}]: dispatch 2026-03-06T10:01:20.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:20 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:20.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:20 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-06T10:01:20.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:20 vm01 ceph-mon[50931]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-06T10:01:20.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:20 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm01.local:9095"}]: dispatch 2026-03-06T10:01:20.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:20 vm01 ceph-mon[50931]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm01.local:9095"}]: dispatch 2026-03-06T10:01:20.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:20 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:20.768 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:20 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T10:01:20.784 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd stat -f json 2026-03-06T10:01:20.868 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:20 vm07 ceph-mon[55141]: Reconfiguring mon.vm07 (monmap changed)... 2026-03-06T10:01:20.868 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:20 vm07 ceph-mon[55141]: Reconfiguring daemon mon.vm07 on vm07 2026-03-06T10:01:20.868 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:20 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/3538377112' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:20.868 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:20 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:20.868 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:20 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:20.868 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:20 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-06T10:01:20.868 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:20 vm07 ceph-mon[55141]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-06T10:01:20.868 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:20 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm01.local:9093"}]: dispatch 2026-03-06T10:01:20.868 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:20 vm07 ceph-mon[55141]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm01.local:9093"}]: dispatch 2026-03-06T10:01:20.868 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:20 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:20.868 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:20 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-06T10:01:20.868 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:20 vm07 ceph-mon[55141]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-06T10:01:20.868 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:20 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm01.local:3000"}]: dispatch 2026-03-06T10:01:20.868 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:20 vm07 ceph-mon[55141]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm01.local:3000"}]: dispatch 2026-03-06T10:01:20.868 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:20 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:20.868 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:20 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-06T10:01:20.868 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:20 vm07 ceph-mon[55141]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-06T10:01:20.868 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:20 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm01.local:9095"}]: dispatch 2026-03-06T10:01:20.868 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:20 vm07 ceph-mon[55141]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm01.local:9095"}]: dispatch 2026-03-06T10:01:20.868 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:20 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:20.868 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:20 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T10:01:21.395 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:01:21.838 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:01:22.003 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":5,"num_osds":0,"num_up_osds":0,"osd_up_since":0,"num_in_osds":0,"osd_in_since":0,"num_remapped_pgs":0} 2026-03-06T10:01:22.043 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:21 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:22.043 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:21 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:22.043 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:21 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:22.043 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:21 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:22.043 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:21 vm01 ceph-mon[50931]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:22.043 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:21 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:22.043 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:21 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:22.043 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:21 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:22.043 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:21 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:22.043 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:21 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:22.043 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:21 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T10:01:22.043 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:21 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:22.044 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:21 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T10:01:22.044 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:21 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T10:01:22.044 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:21 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:22.044 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:21 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T10:01:22.044 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:21 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:22.077 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:21 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:22.077 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:21 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:22.077 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:21 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:22.077 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:21 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:22.077 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:21 vm07 ceph-mon[55141]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:22.077 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:21 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:22.077 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:21 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:22.077 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:21 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:22.077 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:21 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:22.077 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:21 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:22.077 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:21 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T10:01:22.077 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:21 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:22.077 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:21 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T10:01:22.077 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:21 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T10:01:22.077 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:21 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:22.077 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:21 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T10:01:22.077 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:21 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:23.004 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd stat -f json 2026-03-06T10:01:23.123 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:22 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/256395281' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:23.231 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:22 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/256395281' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:23.360 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:01:23.827 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:01:23.990 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":7,"num_osds":2,"num_up_osds":0,"osd_up_since":0,"num_in_osds":2,"osd_in_since":1772787683,"num_remapped_pgs":0} 2026-03-06T10:01:24.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:23 vm01 ceph-mon[50931]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:24.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:23 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/412809967' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "5419104f-76e4-4e2f-9035-ff0ac04e9fed"}]: dispatch 2026-03-06T10:01:24.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:23 vm01 ceph-mon[50931]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "5419104f-76e4-4e2f-9035-ff0ac04e9fed"}]: dispatch 2026-03-06T10:01:24.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:23 vm01 ceph-mon[50931]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "5419104f-76e4-4e2f-9035-ff0ac04e9fed"}]': finished 2026-03-06T10:01:24.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:23 vm01 ceph-mon[50931]: osdmap e6: 1 total, 0 up, 1 in 2026-03-06T10:01:24.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:23 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T10:01:24.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:23 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/795731789' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "f639e6c1-958d-4147-b8c2-bf8bb5af0c1e"}]: dispatch 2026-03-06T10:01:24.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:23 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/795731789' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "f639e6c1-958d-4147-b8c2-bf8bb5af0c1e"}]': finished 2026-03-06T10:01:24.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:23 vm01 ceph-mon[50931]: osdmap e7: 2 total, 0 up, 2 in 2026-03-06T10:01:24.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:23 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T10:01:24.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:23 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T10:01:24.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:23 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/341541868' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:24.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:23 vm07 ceph-mon[55141]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:24.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:23 vm07 ceph-mon[55141]: from='client.? 192.168.123.107:0/412809967' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "5419104f-76e4-4e2f-9035-ff0ac04e9fed"}]: dispatch 2026-03-06T10:01:24.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:23 vm07 ceph-mon[55141]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "5419104f-76e4-4e2f-9035-ff0ac04e9fed"}]: dispatch 2026-03-06T10:01:24.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:23 vm07 ceph-mon[55141]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "5419104f-76e4-4e2f-9035-ff0ac04e9fed"}]': finished 2026-03-06T10:01:24.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:23 vm07 ceph-mon[55141]: osdmap e6: 1 total, 0 up, 1 in 2026-03-06T10:01:24.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:23 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T10:01:24.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:23 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/795731789' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "f639e6c1-958d-4147-b8c2-bf8bb5af0c1e"}]: dispatch 2026-03-06T10:01:24.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:23 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/795731789' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "f639e6c1-958d-4147-b8c2-bf8bb5af0c1e"}]': finished 2026-03-06T10:01:24.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:23 vm07 ceph-mon[55141]: osdmap e7: 2 total, 0 up, 2 in 2026-03-06T10:01:24.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:23 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T10:01:24.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:23 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T10:01:24.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:23 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/341541868' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:24.991 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd stat -f json 2026-03-06T10:01:25.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:24 vm07 ceph-mon[55141]: from='client.? 192.168.123.107:0/2695610227' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T10:01:25.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:24 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/971550754' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T10:01:25.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:24 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/2695610227' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T10:01:25.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:24 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/971550754' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T10:01:25.332 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:01:25.668 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:01:25.844 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":7,"num_osds":2,"num_up_osds":0,"osd_up_since":0,"num_in_osds":2,"osd_in_since":1772787683,"num_remapped_pgs":0} 2026-03-06T10:01:26.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:25 vm01 ceph-mon[50931]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:26.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:25 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/1718403360' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:26.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:25 vm07 ceph-mon[55141]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:26.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:25 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/1718403360' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:26.844 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd stat -f json 2026-03-06T10:01:27.234 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:01:27.592 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:01:27.751 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":9,"num_osds":4,"num_up_osds":0,"osd_up_since":0,"num_in_osds":4,"osd_in_since":1772787687,"num_remapped_pgs":0} 2026-03-06T10:01:27.886 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:27 vm01 ceph-mon[50931]: pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:28.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:27 vm07 ceph-mon[55141]: pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:28.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:27 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/2654123328' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "d645274a-f25c-4360-be0f-460223194536"}]: dispatch 2026-03-06T10:01:28.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:27 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/2654123328' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "d645274a-f25c-4360-be0f-460223194536"}]': finished 2026-03-06T10:01:28.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:27 vm07 ceph-mon[55141]: osdmap e8: 3 total, 0 up, 3 in 2026-03-06T10:01:28.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:27 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T10:01:28.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:27 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T10:01:28.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:27 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:28.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:27 vm07 ceph-mon[55141]: from='client.? 192.168.123.107:0/2049081381' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "75751174-b0ab-4614-a691-06b4cd42e273"}]: dispatch 2026-03-06T10:01:28.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:27 vm07 ceph-mon[55141]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "75751174-b0ab-4614-a691-06b4cd42e273"}]: dispatch 2026-03-06T10:01:28.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:27 vm07 ceph-mon[55141]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "75751174-b0ab-4614-a691-06b4cd42e273"}]': finished 2026-03-06T10:01:28.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:27 vm07 ceph-mon[55141]: osdmap e9: 4 total, 0 up, 4 in 2026-03-06T10:01:28.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:27 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T10:01:28.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:27 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T10:01:28.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:27 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:28.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:27 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T10:01:28.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:27 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/2530785663' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T10:01:28.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:27 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/776551681' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:28.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:27 vm07 ceph-mon[55141]: from='client.? 192.168.123.107:0/2750496136' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T10:01:28.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:27 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/2654123328' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "d645274a-f25c-4360-be0f-460223194536"}]: dispatch 2026-03-06T10:01:28.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:27 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/2654123328' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "d645274a-f25c-4360-be0f-460223194536"}]': finished 2026-03-06T10:01:28.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:27 vm01 ceph-mon[50931]: osdmap e8: 3 total, 0 up, 3 in 2026-03-06T10:01:28.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:27 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T10:01:28.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:27 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T10:01:28.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:27 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:28.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:27 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/2049081381' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "75751174-b0ab-4614-a691-06b4cd42e273"}]: dispatch 2026-03-06T10:01:28.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:27 vm01 ceph-mon[50931]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "75751174-b0ab-4614-a691-06b4cd42e273"}]: dispatch 2026-03-06T10:01:28.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:27 vm01 ceph-mon[50931]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "75751174-b0ab-4614-a691-06b4cd42e273"}]': finished 2026-03-06T10:01:28.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:27 vm01 ceph-mon[50931]: osdmap e9: 4 total, 0 up, 4 in 2026-03-06T10:01:28.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:27 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T10:01:28.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:27 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T10:01:28.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:27 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:28.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:27 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T10:01:28.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:27 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/2530785663' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T10:01:28.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:27 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/776551681' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:28.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:27 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/2750496136' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T10:01:28.752 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd stat -f json 2026-03-06T10:01:29.069 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:01:29.427 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:01:29.596 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":9,"num_osds":4,"num_up_osds":0,"osd_up_since":0,"num_in_osds":4,"osd_in_since":1772787687,"num_remapped_pgs":0} 2026-03-06T10:01:30.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:29 vm07 ceph-mon[55141]: pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:30.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:29 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/3773979978' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:30.220 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:29 vm01 ceph-mon[50931]: pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:30.220 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:29 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/3773979978' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:30.597 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd stat -f json 2026-03-06T10:01:30.960 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:01:31.117 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:30 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/1204719557' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "84dec1a3-932d-4894-b752-0323fbc97b37"}]: dispatch 2026-03-06T10:01:31.117 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:30 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/1204719557' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "84dec1a3-932d-4894-b752-0323fbc97b37"}]': finished 2026-03-06T10:01:31.117 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:30 vm07 ceph-mon[55141]: osdmap e10: 5 total, 0 up, 5 in 2026-03-06T10:01:31.117 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:30 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T10:01:31.117 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:30 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T10:01:31.117 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:30 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:31.117 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:30 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T10:01:31.117 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:30 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:31.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:30 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/1204719557' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "84dec1a3-932d-4894-b752-0323fbc97b37"}]: dispatch 2026-03-06T10:01:31.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:30 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/1204719557' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "84dec1a3-932d-4894-b752-0323fbc97b37"}]': finished 2026-03-06T10:01:31.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:30 vm01 ceph-mon[50931]: osdmap e10: 5 total, 0 up, 5 in 2026-03-06T10:01:31.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:30 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T10:01:31.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:30 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T10:01:31.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:30 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:31.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:30 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T10:01:31.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:30 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:31.340 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:01:31.490 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":11,"num_osds":6,"num_up_osds":0,"osd_up_since":0,"num_in_osds":6,"osd_in_since":1772787691,"num_remapped_pgs":0} 2026-03-06T10:01:31.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:31 vm07 ceph-mon[55141]: pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:31.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:31 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/927715895' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T10:01:31.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:31 vm07 ceph-mon[55141]: from='client.? 192.168.123.107:0/3195265357' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "036ff726-417f-464a-b30b-e5d7e55a059f"}]: dispatch 2026-03-06T10:01:31.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:31 vm07 ceph-mon[55141]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "036ff726-417f-464a-b30b-e5d7e55a059f"}]: dispatch 2026-03-06T10:01:31.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:31 vm07 ceph-mon[55141]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "036ff726-417f-464a-b30b-e5d7e55a059f"}]': finished 2026-03-06T10:01:31.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:31 vm07 ceph-mon[55141]: osdmap e11: 6 total, 0 up, 6 in 2026-03-06T10:01:31.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:31 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T10:01:31.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:31 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T10:01:31.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:31 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:31.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:31 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T10:01:31.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:31 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:31.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:31 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:31.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:31 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/1498336243' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:31.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:31 vm07 ceph-mon[55141]: from='client.? 192.168.123.107:0/1419073911' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T10:01:32.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:31 vm01 ceph-mon[50931]: pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:32.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:31 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/927715895' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T10:01:32.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:31 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/3195265357' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "036ff726-417f-464a-b30b-e5d7e55a059f"}]: dispatch 2026-03-06T10:01:32.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:31 vm01 ceph-mon[50931]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "036ff726-417f-464a-b30b-e5d7e55a059f"}]: dispatch 2026-03-06T10:01:32.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:31 vm01 ceph-mon[50931]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "036ff726-417f-464a-b30b-e5d7e55a059f"}]': finished 2026-03-06T10:01:32.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:31 vm01 ceph-mon[50931]: osdmap e11: 6 total, 0 up, 6 in 2026-03-06T10:01:32.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:31 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T10:01:32.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:31 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T10:01:32.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:31 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:32.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:31 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T10:01:32.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:31 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:32.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:31 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:32.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:31 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/1498336243' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:32.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:31 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/1419073911' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T10:01:32.491 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd stat -f json 2026-03-06T10:01:32.824 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:01:33.153 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:01:33.301 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":11,"num_osds":6,"num_up_osds":0,"osd_up_since":0,"num_in_osds":6,"osd_in_since":1772787691,"num_remapped_pgs":0} 2026-03-06T10:01:34.187 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:33 vm07 ceph-mon[55141]: pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:34.187 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:33 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/1419132069' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:34.187 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:33 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T10:01:34.194 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:33 vm01 ceph-mon[50931]: pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:34.194 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:33 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/1419132069' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:34.194 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:33 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T10:01:34.303 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd stat -f json 2026-03-06T10:01:34.674 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:01:35.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:34 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/2872678081' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "1705c5f8-0f21-4e35-a9df-ef5ae49cc4c3"}]: dispatch 2026-03-06T10:01:35.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:34 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/2872678081' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "1705c5f8-0f21-4e35-a9df-ef5ae49cc4c3"}]': finished 2026-03-06T10:01:35.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:34 vm01 ceph-mon[50931]: osdmap e12: 7 total, 0 up, 7 in 2026-03-06T10:01:35.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:34 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T10:01:35.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:34 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T10:01:35.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:34 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:35.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:34 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T10:01:35.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:34 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:35.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:34 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:35.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:34 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:35.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:34 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/4213663534' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T10:01:35.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:34 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/2594233750' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e96c75da-b3cf-40d0-930b-2665d7b20d97"}]: dispatch 2026-03-06T10:01:35.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:34 vm01 ceph-mon[50931]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e96c75da-b3cf-40d0-930b-2665d7b20d97"}]: dispatch 2026-03-06T10:01:35.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:34 vm01 ceph-mon[50931]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "e96c75da-b3cf-40d0-930b-2665d7b20d97"}]': finished 2026-03-06T10:01:35.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:34 vm01 ceph-mon[50931]: osdmap e13: 8 total, 0 up, 8 in 2026-03-06T10:01:35.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:34 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T10:01:35.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:34 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T10:01:35.019 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:34 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:35.019 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:34 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T10:01:35.019 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:34 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:35.019 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:34 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:35.019 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:34 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:35.019 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:34 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:35.031 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:01:35.141 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:34 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/2872678081' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "1705c5f8-0f21-4e35-a9df-ef5ae49cc4c3"}]: dispatch 2026-03-06T10:01:35.141 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:34 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/2872678081' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "1705c5f8-0f21-4e35-a9df-ef5ae49cc4c3"}]': finished 2026-03-06T10:01:35.141 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:34 vm07 ceph-mon[55141]: osdmap e12: 7 total, 0 up, 7 in 2026-03-06T10:01:35.141 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:34 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T10:01:35.141 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:34 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T10:01:35.141 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:34 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:35.141 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:34 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T10:01:35.141 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:34 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:35.141 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:34 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:35.141 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:34 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:35.141 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:34 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/4213663534' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T10:01:35.141 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:34 vm07 ceph-mon[55141]: from='client.? 192.168.123.107:0/2594233750' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e96c75da-b3cf-40d0-930b-2665d7b20d97"}]: dispatch 2026-03-06T10:01:35.141 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:34 vm07 ceph-mon[55141]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e96c75da-b3cf-40d0-930b-2665d7b20d97"}]: dispatch 2026-03-06T10:01:35.141 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:34 vm07 ceph-mon[55141]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "e96c75da-b3cf-40d0-930b-2665d7b20d97"}]': finished 2026-03-06T10:01:35.141 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:34 vm07 ceph-mon[55141]: osdmap e13: 8 total, 0 up, 8 in 2026-03-06T10:01:35.141 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:34 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T10:01:35.141 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:34 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T10:01:35.141 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:34 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:35.141 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:34 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T10:01:35.141 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:34 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:35.141 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:34 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:35.141 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:34 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:35.141 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:34 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:35.201 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":13,"num_osds":8,"num_up_osds":0,"osd_up_since":0,"num_in_osds":8,"osd_in_since":1772787694,"num_remapped_pgs":0} 2026-03-06T10:01:36.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:35 vm07 ceph-mon[55141]: pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:36.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:35 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/4130661957' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:36.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:35 vm07 ceph-mon[55141]: from='client.? 192.168.123.107:0/1685953925' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T10:01:36.202 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd stat -f json 2026-03-06T10:01:36.225 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:35 vm01 ceph-mon[50931]: pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:36.225 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:35 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/4130661957' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:36.225 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:35 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/1685953925' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T10:01:36.544 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:01:36.887 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:01:37.068 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":13,"num_osds":8,"num_up_osds":0,"osd_up_since":0,"num_in_osds":8,"osd_in_since":1772787694,"num_remapped_pgs":0} 2026-03-06T10:01:37.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:36 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/2082239135' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:37.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:36 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/2082239135' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:38.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:37 vm01 ceph-mon[50931]: pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:38.069 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd stat -f json 2026-03-06T10:01:38.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:37 vm07 ceph-mon[55141]: pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:38.498 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:01:39.084 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:01:39.240 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":13,"num_osds":8,"num_up_osds":0,"osd_up_since":0,"num_in_osds":8,"osd_in_since":1772787694,"num_remapped_pgs":0} 2026-03-06T10:01:40.022 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:39 vm01 ceph-mon[50931]: pgmap v26: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:40.022 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:39 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/3498029537' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:40.022 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:39 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-06T10:01:40.022 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:39 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:40.022 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:39 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-06T10:01:40.022 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:39 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:40.168 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:39 vm07 ceph-mon[55141]: pgmap v26: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:40.168 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:39 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/3498029537' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:40.168 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:39 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-06T10:01:40.169 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:39 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:40.169 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:39 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-06T10:01:40.169 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:39 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:40.241 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd stat -f json 2026-03-06T10:01:40.784 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:01:40.985 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:40 vm01 ceph-mon[50931]: Deploying daemon osd.1 on vm01 2026-03-06T10:01:40.985 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:40 vm01 ceph-mon[50931]: Deploying daemon osd.0 on vm07 2026-03-06T10:01:41.144 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:40 vm07 ceph-mon[55141]: Deploying daemon osd.1 on vm01 2026-03-06T10:01:41.144 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:40 vm07 ceph-mon[55141]: Deploying daemon osd.0 on vm07 2026-03-06T10:01:41.288 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:01:41.490 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":13,"num_osds":8,"num_up_osds":0,"osd_up_since":0,"num_in_osds":8,"osd_in_since":1772787694,"num_remapped_pgs":0} 2026-03-06T10:01:42.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:41 vm01 ceph-mon[50931]: pgmap v27: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:42.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:41 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/3137768209' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:42.428 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:41 vm07 ceph-mon[55141]: pgmap v27: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:42.429 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:41 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/3137768209' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:42.491 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd stat -f json 2026-03-06T10:01:42.927 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:01:43.311 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:01:43.498 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":13,"num_osds":8,"num_up_osds":0,"osd_up_since":0,"num_in_osds":8,"osd_in_since":1772787694,"num_remapped_pgs":0} 2026-03-06T10:01:44.177 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:43 vm07 ceph-mon[55141]: pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:44.177 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:43 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:44.177 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:43 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:44.177 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:43 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-06T10:01:44.177 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:43 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:44.177 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:43 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/1130922686' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:44.177 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:43 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:44.177 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:43 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:44.177 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:43 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-06T10:01:44.177 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:43 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:44.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:43 vm01 ceph-mon[50931]: pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:44.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:43 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:44.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:43 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:44.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:43 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-06T10:01:44.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:43 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:44.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:43 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/1130922686' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:44.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:43 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:44.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:43 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:44.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:43 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-06T10:01:44.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:43 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:44.499 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd stat -f json 2026-03-06T10:01:44.913 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:01:45.189 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:44 vm07 ceph-mon[55141]: Deploying daemon osd.3 on vm07 2026-03-06T10:01:45.190 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:44 vm07 ceph-mon[55141]: Deploying daemon osd.2 on vm01 2026-03-06T10:01:45.190 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:44 vm07 ceph-mon[55141]: from='osd.0 [v2:192.168.123.107:6800/4015024690,v1:192.168.123.107:6801/4015024690]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-06T10:01:45.190 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:44 vm07 ceph-mon[55141]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-06T10:01:45.240 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:44 vm01 ceph-mon[50931]: Deploying daemon osd.3 on vm07 2026-03-06T10:01:45.240 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:44 vm01 ceph-mon[50931]: Deploying daemon osd.2 on vm01 2026-03-06T10:01:45.240 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:44 vm01 ceph-mon[50931]: from='osd.0 [v2:192.168.123.107:6800/4015024690,v1:192.168.123.107:6801/4015024690]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-06T10:01:45.240 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:44 vm01 ceph-mon[50931]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-06T10:01:45.507 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:01:45.671 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":14,"num_osds":8,"num_up_osds":0,"osd_up_since":0,"num_in_osds":8,"osd_in_since":1772787694,"num_remapped_pgs":0} 2026-03-06T10:01:46.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:46 vm07 ceph-mon[55141]: pgmap v29: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:46.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:46 vm07 ceph-mon[55141]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-06T10:01:46.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:46 vm07 ceph-mon[55141]: osdmap e14: 8 total, 0 up, 8 in 2026-03-06T10:01:46.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:46 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T10:01:46.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:46 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T10:01:46.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:46 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:46.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:46 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T10:01:46.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:46 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:46.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:46 vm07 ceph-mon[55141]: from='osd.0 [v2:192.168.123.107:6800/4015024690,v1:192.168.123.107:6801/4015024690]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-06T10:01:46.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:46 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:46.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:46 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:46.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:46 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:46.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:46 vm07 ceph-mon[55141]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-06T10:01:46.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:46 vm07 ceph-mon[55141]: from='osd.1 [v2:192.168.123.101:6802/3951729305,v1:192.168.123.101:6803/3951729305]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-06T10:01:46.197 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:46 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/2506124188' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:46.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:46 vm01 ceph-mon[50931]: pgmap v29: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:46.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:46 vm01 ceph-mon[50931]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-06T10:01:46.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:46 vm01 ceph-mon[50931]: osdmap e14: 8 total, 0 up, 8 in 2026-03-06T10:01:46.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:46 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T10:01:46.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:46 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T10:01:46.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:46 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:46.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:46 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T10:01:46.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:46 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:46.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:46 vm01 ceph-mon[50931]: from='osd.0 [v2:192.168.123.107:6800/4015024690,v1:192.168.123.107:6801/4015024690]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-06T10:01:46.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:46 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:46.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:46 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:46.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:46 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:46.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:46 vm01 ceph-mon[50931]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-06T10:01:46.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:46 vm01 ceph-mon[50931]: from='osd.1 [v2:192.168.123.101:6802/3951729305,v1:192.168.123.101:6803/3951729305]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-06T10:01:46.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:46 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/2506124188' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:46.672 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd stat -f json 2026-03-06T10:01:47.049 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:01:47.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-06T10:01:47.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: from='osd.1 [v2:192.168.123.101:6802/3951729305,v1:192.168.123.101:6803/3951729305]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-06T10:01:47.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: osdmap e15: 8 total, 0 up, 8 in 2026-03-06T10:01:47.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: from='osd.1 [v2:192.168.123.101:6802/3951729305,v1:192.168.123.101:6803/3951729305]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-06T10:01:47.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T10:01:47.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T10:01:47.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:47.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T10:01:47.269 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:47.269 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:47.269 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:47.269 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:47.269 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T10:01:47.269 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:47.269 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:47.269 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-06T10:01:47.269 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:47.269 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: from='osd.0 ' entity='osd.0' 2026-03-06T10:01:47.269 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: from='osd.1 [v2:192.168.123.101:6802/3951729305,v1:192.168.123.101:6803/3951729305]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-06T10:01:47.269 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: osd.0 [v2:192.168.123.107:6800/4015024690,v1:192.168.123.107:6801/4015024690] boot 2026-03-06T10:01:47.269 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: osdmap e16: 8 total, 1 up, 8 in 2026-03-06T10:01:47.269 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T10:01:47.269 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T10:01:47.269 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:47.269 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T10:01:47.269 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:47.269 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:47.269 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:47.269 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:47.269 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:47 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T10:01:47.274 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-06T10:01:47.274 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: from='osd.1 [v2:192.168.123.101:6802/3951729305,v1:192.168.123.101:6803/3951729305]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-06T10:01:47.274 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: osdmap e15: 8 total, 0 up, 8 in 2026-03-06T10:01:47.274 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: from='osd.1 [v2:192.168.123.101:6802/3951729305,v1:192.168.123.101:6803/3951729305]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-06T10:01:47.274 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T10:01:47.274 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T10:01:47.274 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:47.274 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T10:01:47.275 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:47.275 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:47.275 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:47.275 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:47.275 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T10:01:47.275 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:47.275 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:47.275 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-06T10:01:47.275 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:47.275 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: from='osd.0 ' entity='osd.0' 2026-03-06T10:01:47.275 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: from='osd.1 [v2:192.168.123.101:6802/3951729305,v1:192.168.123.101:6803/3951729305]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-06T10:01:47.275 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: osd.0 [v2:192.168.123.107:6800/4015024690,v1:192.168.123.107:6801/4015024690] boot 2026-03-06T10:01:47.275 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: osdmap e16: 8 total, 1 up, 8 in 2026-03-06T10:01:47.275 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T10:01:47.275 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T10:01:47.275 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:47.275 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T10:01:47.275 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:47.275 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:47.275 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:47.275 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:47.275 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:47 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T10:01:47.424 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:01:47.724 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":16,"num_osds":8,"num_up_osds":1,"osd_up_since":1772787706,"num_in_osds":8,"osd_in_since":1772787694,"num_remapped_pgs":0} 2026-03-06T10:01:48.254 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:48 vm07 ceph-mon[55141]: purged_snaps scrub starts 2026-03-06T10:01:48.254 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:48 vm07 ceph-mon[55141]: purged_snaps scrub ok 2026-03-06T10:01:48.254 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:48 vm07 ceph-mon[55141]: Deploying daemon osd.5 on vm07 2026-03-06T10:01:48.254 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:48 vm07 ceph-mon[55141]: pgmap v32: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:48.254 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:48 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/2783586696' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:48.254 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:48 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:48.254 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:48 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:48.254 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:48 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-06T10:01:48.254 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:48 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:48.254 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:48 vm07 ceph-mon[55141]: osd.1 [v2:192.168.123.101:6802/3951729305,v1:192.168.123.101:6803/3951729305] boot 2026-03-06T10:01:48.254 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:48 vm07 ceph-mon[55141]: osdmap e17: 8 total, 2 up, 8 in 2026-03-06T10:01:48.254 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:48 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T10:01:48.254 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:48 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:48.254 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:48 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T10:01:48.254 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:48 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:48.254 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:48 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:48.254 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:48 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:48.254 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:48 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:48.377 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:48 vm01 ceph-mon[50931]: purged_snaps scrub starts 2026-03-06T10:01:48.377 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:48 vm01 ceph-mon[50931]: purged_snaps scrub ok 2026-03-06T10:01:48.377 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:48 vm01 ceph-mon[50931]: Deploying daemon osd.5 on vm07 2026-03-06T10:01:48.377 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:48 vm01 ceph-mon[50931]: pgmap v32: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T10:01:48.377 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:48 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/2783586696' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:48.377 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:48 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:48.377 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:48 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:48.377 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:48 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-06T10:01:48.377 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:48 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:48.377 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:48 vm01 ceph-mon[50931]: osd.1 [v2:192.168.123.101:6802/3951729305,v1:192.168.123.101:6803/3951729305] boot 2026-03-06T10:01:48.377 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:48 vm01 ceph-mon[50931]: osdmap e17: 8 total, 2 up, 8 in 2026-03-06T10:01:48.377 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:48 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T10:01:48.377 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:48 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:48.377 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:48 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T10:01:48.377 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:48 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:48.377 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:48 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:48.377 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:48 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:48.377 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:48 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:48.725 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd stat -f json 2026-03-06T10:01:49.143 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:01:49.343 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:49 vm01 ceph-mon[50931]: purged_snaps scrub starts 2026-03-06T10:01:49.343 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:49 vm01 ceph-mon[50931]: purged_snaps scrub ok 2026-03-06T10:01:49.343 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:49 vm01 ceph-mon[50931]: Deploying daemon osd.4 on vm01 2026-03-06T10:01:49.343 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:49 vm01 ceph-mon[50931]: from='osd.3 [v2:192.168.123.107:6808/206187849,v1:192.168.123.107:6809/206187849]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-06T10:01:49.343 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:49 vm01 ceph-mon[50931]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-06T10:01:49.343 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:49 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T10:01:49.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:49 vm07 ceph-mon[55141]: purged_snaps scrub starts 2026-03-06T10:01:49.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:49 vm07 ceph-mon[55141]: purged_snaps scrub ok 2026-03-06T10:01:49.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:49 vm07 ceph-mon[55141]: Deploying daemon osd.4 on vm01 2026-03-06T10:01:49.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:49 vm07 ceph-mon[55141]: from='osd.3 [v2:192.168.123.107:6808/206187849,v1:192.168.123.107:6809/206187849]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-06T10:01:49.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:49 vm07 ceph-mon[55141]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-06T10:01:49.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:49 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T10:01:49.886 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:01:50.106 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":18,"num_osds":8,"num_up_osds":2,"osd_up_since":1772787708,"num_in_osds":8,"osd_in_since":1772787694,"num_remapped_pgs":0} 2026-03-06T10:01:50.164 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:50 vm07 ceph-mon[55141]: pgmap v35: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T10:01:50.164 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:50 vm07 ceph-mon[55141]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-06T10:01:50.164 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:50 vm07 ceph-mon[55141]: osdmap e18: 8 total, 2 up, 8 in 2026-03-06T10:01:50.164 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:50 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:50.164 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:50 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T10:01:50.164 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:50 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:50.164 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:50 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:50.164 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:50 vm07 ceph-mon[55141]: from='osd.3 [v2:192.168.123.107:6808/206187849,v1:192.168.123.107:6809/206187849]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-06T10:01:50.164 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:50 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:50.164 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:50 vm07 ceph-mon[55141]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-06T10:01:50.164 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:50 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:50.164 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:50 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/1347617822' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:50.164 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:50 vm07 ceph-mon[55141]: from='osd.2 [v2:192.168.123.101:6810/970532629,v1:192.168.123.101:6811/970532629]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-06T10:01:50.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:50 vm01 ceph-mon[50931]: pgmap v35: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T10:01:50.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:50 vm01 ceph-mon[50931]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-06T10:01:50.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:50 vm01 ceph-mon[50931]: osdmap e18: 8 total, 2 up, 8 in 2026-03-06T10:01:50.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:50 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:50.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:50 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T10:01:50.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:50 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:50.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:50 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:50.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:50 vm01 ceph-mon[50931]: from='osd.3 [v2:192.168.123.107:6808/206187849,v1:192.168.123.107:6809/206187849]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-06T10:01:50.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:50 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:50.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:50 vm01 ceph-mon[50931]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-06T10:01:50.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:50 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:50.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:50 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/1347617822' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:50.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:50 vm01 ceph-mon[50931]: from='osd.2 [v2:192.168.123.101:6810/970532629,v1:192.168.123.101:6811/970532629]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-06T10:01:51.107 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd stat -f json 2026-03-06T10:01:51.169 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:51 vm07 ceph-mon[55141]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-06T10:01:51.169 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:51 vm07 ceph-mon[55141]: from='osd.2 [v2:192.168.123.101:6810/970532629,v1:192.168.123.101:6811/970532629]' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-06T10:01:51.169 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:51 vm07 ceph-mon[55141]: osdmap e19: 8 total, 2 up, 8 in 2026-03-06T10:01:51.169 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:51 vm07 ceph-mon[55141]: from='osd.2 [v2:192.168.123.101:6810/970532629,v1:192.168.123.101:6811/970532629]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-06T10:01:51.169 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:51 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:51.169 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:51 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T10:01:51.169 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:51 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:51.169 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:51 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:51.169 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:51 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:51.169 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:51 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:51.169 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:51 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T10:01:51.169 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:51 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:51.169 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:51 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:51.169 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:51 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-06T10:01:51.169 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:51 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:51.169 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:51 vm07 ceph-mon[55141]: Deploying daemon osd.7 on vm07 2026-03-06T10:01:51.169 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:51 vm07 ceph-mon[55141]: pgmap v38: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T10:01:51.169 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:51 vm07 ceph-mon[55141]: from='osd.2 [v2:192.168.123.101:6810/970532629,v1:192.168.123.101:6811/970532629]' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-06T10:01:51.169 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:51 vm07 ceph-mon[55141]: osd.3 [v2:192.168.123.107:6808/206187849,v1:192.168.123.107:6809/206187849] boot 2026-03-06T10:01:51.169 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:51 vm07 ceph-mon[55141]: osdmap e20: 8 total, 3 up, 8 in 2026-03-06T10:01:51.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:51 vm01 ceph-mon[50931]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-06T10:01:51.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:51 vm01 ceph-mon[50931]: from='osd.2 [v2:192.168.123.101:6810/970532629,v1:192.168.123.101:6811/970532629]' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-06T10:01:51.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:51 vm01 ceph-mon[50931]: osdmap e19: 8 total, 2 up, 8 in 2026-03-06T10:01:51.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:51 vm01 ceph-mon[50931]: from='osd.2 [v2:192.168.123.101:6810/970532629,v1:192.168.123.101:6811/970532629]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-06T10:01:51.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:51 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:51.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:51 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T10:01:51.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:51 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:51.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:51 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:51.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:51 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:51.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:51 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:51.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:51 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T10:01:51.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:51 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:51.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:51 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:51.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:51 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-06T10:01:51.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:51 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:51.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:51 vm01 ceph-mon[50931]: Deploying daemon osd.7 on vm07 2026-03-06T10:01:51.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:51 vm01 ceph-mon[50931]: pgmap v38: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T10:01:51.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:51 vm01 ceph-mon[50931]: from='osd.2 [v2:192.168.123.101:6810/970532629,v1:192.168.123.101:6811/970532629]' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-06T10:01:51.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:51 vm01 ceph-mon[50931]: osd.3 [v2:192.168.123.107:6808/206187849,v1:192.168.123.107:6809/206187849] boot 2026-03-06T10:01:51.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:51 vm01 ceph-mon[50931]: osdmap e20: 8 total, 3 up, 8 in 2026-03-06T10:01:51.604 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:01:52.051 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:01:52.232 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":20,"num_osds":8,"num_up_osds":3,"osd_up_since":1772787711,"num_in_osds":8,"osd_in_since":1772787694,"num_remapped_pgs":0} 2026-03-06T10:01:52.245 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:52 vm07 ceph-mon[55141]: purged_snaps scrub starts 2026-03-06T10:01:52.245 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:52 vm07 ceph-mon[55141]: purged_snaps scrub ok 2026-03-06T10:01:52.245 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:52 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:52.245 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:52 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T10:01:52.245 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:52 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:52.245 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:52 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:52.245 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:52 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:52.245 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:52 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:52.245 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:52 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:52.245 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:52 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:52.245 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:52 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:52.245 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:52 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-06T10:01:52.245 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:52 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:52.245 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:52 vm07 ceph-mon[55141]: Deploying daemon osd.6 on vm01 2026-03-06T10:01:52.245 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:52 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/3200790495' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:52.245 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:52 vm07 ceph-mon[55141]: osd.2 [v2:192.168.123.101:6810/970532629,v1:192.168.123.101:6811/970532629] boot 2026-03-06T10:01:52.245 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:52 vm07 ceph-mon[55141]: osdmap e21: 8 total, 4 up, 8 in 2026-03-06T10:01:52.245 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:52 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:52.245 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:52 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:52.245 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:52 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:52.245 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:52 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:52.245 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:52 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:52.302 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:52 vm01 ceph-mon[50931]: purged_snaps scrub starts 2026-03-06T10:01:52.302 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:52 vm01 ceph-mon[50931]: purged_snaps scrub ok 2026-03-06T10:01:52.302 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:52 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:52.302 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:52 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T10:01:52.302 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:52 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:52.302 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:52 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:52.302 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:52 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:52.302 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:52 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:52.302 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:52 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:52.302 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:52 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:52.302 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:52 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:52.302 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:52 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-06T10:01:52.302 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:52 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:01:52.302 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:52 vm01 ceph-mon[50931]: Deploying daemon osd.6 on vm01 2026-03-06T10:01:52.302 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:52 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/3200790495' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:52.302 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:52 vm01 ceph-mon[50931]: osd.2 [v2:192.168.123.101:6810/970532629,v1:192.168.123.101:6811/970532629] boot 2026-03-06T10:01:52.302 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:52 vm01 ceph-mon[50931]: osdmap e21: 8 total, 4 up, 8 in 2026-03-06T10:01:52.302 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:52 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T10:01:52.302 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:52 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:52.303 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:52 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:52.303 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:52 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:52.303 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:52 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:53.234 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd stat -f json 2026-03-06T10:01:53.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:53 vm01 ceph-mon[50931]: purged_snaps scrub starts 2026-03-06T10:01:53.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:53 vm01 ceph-mon[50931]: purged_snaps scrub ok 2026-03-06T10:01:53.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:53 vm01 ceph-mon[50931]: from='osd.5 [v2:192.168.123.107:6816/1614548926,v1:192.168.123.107:6817/1614548926]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-06T10:01:53.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:53 vm01 ceph-mon[50931]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-06T10:01:53.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:53 vm01 ceph-mon[50931]: pgmap v41: 0 pgs: ; 0 B data, 105 MiB used, 80 GiB / 80 GiB avail 2026-03-06T10:01:53.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:53 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-06T10:01:53.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:53 vm01 ceph-mon[50931]: from='osd.4 [v2:192.168.123.101:6818/3295130738,v1:192.168.123.101:6819/3295130738]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-06T10:01:53.367 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:53 vm07 ceph-mon[55141]: purged_snaps scrub starts 2026-03-06T10:01:53.367 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:53 vm07 ceph-mon[55141]: purged_snaps scrub ok 2026-03-06T10:01:53.367 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:53 vm07 ceph-mon[55141]: from='osd.5 [v2:192.168.123.107:6816/1614548926,v1:192.168.123.107:6817/1614548926]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-06T10:01:53.367 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:53 vm07 ceph-mon[55141]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-06T10:01:53.367 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:53 vm07 ceph-mon[55141]: pgmap v41: 0 pgs: ; 0 B data, 105 MiB used, 80 GiB / 80 GiB avail 2026-03-06T10:01:53.367 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:53 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-06T10:01:53.367 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:53 vm07 ceph-mon[55141]: from='osd.4 [v2:192.168.123.101:6818/3295130738,v1:192.168.123.101:6819/3295130738]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-06T10:01:53.786 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:01:54.149 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:54 vm01 ceph-mon[50931]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-06T10:01:54.150 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:54 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-06T10:01:54.150 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:54 vm01 ceph-mon[50931]: from='osd.4 [v2:192.168.123.101:6818/3295130738,v1:192.168.123.101:6819/3295130738]' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-06T10:01:54.150 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:54 vm01 ceph-mon[50931]: osdmap e22: 8 total, 4 up, 8 in 2026-03-06T10:01:54.150 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:54 vm01 ceph-mon[50931]: from='osd.4 [v2:192.168.123.101:6818/3295130738,v1:192.168.123.101:6819/3295130738]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-06T10:01:54.150 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:54 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:54.150 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:54 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:54.150 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:54 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:54.150 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:54 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:54.150 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:54 vm01 ceph-mon[50931]: from='osd.5 [v2:192.168.123.107:6816/1614548926,v1:192.168.123.107:6817/1614548926]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-06T10:01:54.150 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:54 vm01 ceph-mon[50931]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-06T10:01:54.150 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:54 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-06T10:01:54.150 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:54 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:54.150 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:54 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:54.173 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:01:54.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:54 vm07 ceph-mon[55141]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-06T10:01:54.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:54 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-06T10:01:54.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:54 vm07 ceph-mon[55141]: from='osd.4 [v2:192.168.123.101:6818/3295130738,v1:192.168.123.101:6819/3295130738]' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-06T10:01:54.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:54 vm07 ceph-mon[55141]: osdmap e22: 8 total, 4 up, 8 in 2026-03-06T10:01:54.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:54 vm07 ceph-mon[55141]: from='osd.4 [v2:192.168.123.101:6818/3295130738,v1:192.168.123.101:6819/3295130738]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-06T10:01:54.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:54 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:54.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:54 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:54.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:54 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:54.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:54 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:54.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:54 vm07 ceph-mon[55141]: from='osd.5 [v2:192.168.123.107:6816/1614548926,v1:192.168.123.107:6817/1614548926]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-06T10:01:54.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:54 vm07 ceph-mon[55141]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-06T10:01:54.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:54 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-06T10:01:54.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:54 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:54.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:54 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:54.607 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":23,"num_osds":8,"num_up_osds":4,"osd_up_since":1772787712,"num_in_osds":8,"osd_in_since":1772787694,"num_remapped_pgs":0} 2026-03-06T10:01:55.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:54 vm01 sudo[77597]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-06T10:01:55.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:54 vm01 sudo[77597]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-06T10:01:55.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:54 vm01 sudo[77597]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-06T10:01:55.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:54 vm01 sudo[77597]: pam_unix(sudo:session): session closed for user root 2026-03-06T10:01:55.134 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:54 vm07 sudo[65929]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-06T10:01:55.134 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:54 vm07 sudo[65929]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-06T10:01:55.134 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:54 vm07 sudo[65929]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-06T10:01:55.134 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:54 vm07 sudo[65929]: pam_unix(sudo:session): session closed for user root 2026-03-06T10:01:55.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:55 vm07 ceph-mon[55141]: from='osd.4 [v2:192.168.123.101:6818/3295130738,v1:192.168.123.101:6819/3295130738]' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-06T10:01:55.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:55 vm07 ceph-mon[55141]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-06T10:01:55.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:55 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-06T10:01:55.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:55 vm07 ceph-mon[55141]: osdmap e23: 8 total, 4 up, 8 in 2026-03-06T10:01:55.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:55 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:55.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:55 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:55.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:55 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:55.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:55 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:55.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:55 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:55.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:55 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:55.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:55 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/153840108' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:55.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:55 vm07 ceph-mon[55141]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T10:01:55.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:55 vm07 ceph-mon[55141]: pgmap v44: 1 pgs: 1 unknown; 0 B data, 105 MiB used, 80 GiB / 80 GiB avail 2026-03-06T10:01:55.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:55 vm07 ceph-mon[55141]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T10:01:55.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:55 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-06T10:01:55.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:55 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-06T10:01:55.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:55 vm07 ceph-mon[55141]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T10:01:55.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:55 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-06T10:01:55.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:55 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-06T10:01:55.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:55 vm07 ceph-mon[55141]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T10:01:55.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:55 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:55.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:55 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:55.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:55 vm01 ceph-mon[50931]: from='osd.4 [v2:192.168.123.101:6818/3295130738,v1:192.168.123.101:6819/3295130738]' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-06T10:01:55.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:55 vm01 ceph-mon[50931]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-06T10:01:55.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:55 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-06T10:01:55.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:55 vm01 ceph-mon[50931]: osdmap e23: 8 total, 4 up, 8 in 2026-03-06T10:01:55.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:55 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:55.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:55 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:55.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:55 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:55.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:55 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:55.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:55 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:55.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:55 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:55.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:55 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/153840108' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:55.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:55 vm01 ceph-mon[50931]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T10:01:55.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:55 vm01 ceph-mon[50931]: pgmap v44: 1 pgs: 1 unknown; 0 B data, 105 MiB used, 80 GiB / 80 GiB avail 2026-03-06T10:01:55.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:55 vm01 ceph-mon[50931]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T10:01:55.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:55 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-06T10:01:55.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:55 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-06T10:01:55.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:55 vm01 ceph-mon[50931]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T10:01:55.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:55 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mon metadata", "id": "vm01"}]: dispatch 2026-03-06T10:01:55.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:55 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-06T10:01:55.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:55 vm01 ceph-mon[50931]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T10:01:55.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:55 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:55.487 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:55 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:55.608 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd stat -f json 2026-03-06T10:01:56.023 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:01:56.140 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:56 vm07 ceph-mon[55141]: purged_snaps scrub starts 2026-03-06T10:01:56.140 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:56 vm07 ceph-mon[55141]: purged_snaps scrub ok 2026-03-06T10:01:56.140 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:56 vm07 ceph-mon[55141]: purged_snaps scrub starts 2026-03-06T10:01:56.382 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:56 vm01 ceph-mon[50931]: purged_snaps scrub starts 2026-03-06T10:01:56.382 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:56 vm01 ceph-mon[50931]: purged_snaps scrub ok 2026-03-06T10:01:56.382 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:56 vm01 ceph-mon[50931]: purged_snaps scrub starts 2026-03-06T10:01:56.382 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:56 vm01 ceph-mon[50931]: purged_snaps scrub ok 2026-03-06T10:01:56.382 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:56 vm01 ceph-mon[50931]: mgrmap e19: vm01.ifwqbh(active, since 66s), standbys: vm07.myglqt 2026-03-06T10:01:56.382 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:56 vm01 ceph-mon[50931]: osdmap e24: 8 total, 4 up, 8 in 2026-03-06T10:01:56.382 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:56 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:56.382 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:56 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:56.382 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:56 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:56.382 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:56 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:56.382 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:56 vm01 ceph-mon[50931]: from='osd.4 [v2:192.168.123.101:6818/3295130738,v1:192.168.123.101:6819/3295130738]' entity='osd.4' 2026-03-06T10:01:56.382 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:56 vm01 ceph-mon[50931]: from='osd.5 ' entity='osd.5' 2026-03-06T10:01:56.382 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:56 vm01 ceph-mon[50931]: from='osd.7 [v2:192.168.123.107:6824/1097411483,v1:192.168.123.107:6825/1097411483]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-06T10:01:56.382 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:56 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:56.382 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:56 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:56.382 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:56 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:56.382 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:56 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:56.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:56 vm07 ceph-mon[55141]: purged_snaps scrub ok 2026-03-06T10:01:56.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:56 vm07 ceph-mon[55141]: mgrmap e19: vm01.ifwqbh(active, since 66s), standbys: vm07.myglqt 2026-03-06T10:01:56.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:56 vm07 ceph-mon[55141]: osdmap e24: 8 total, 4 up, 8 in 2026-03-06T10:01:56.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:56 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:56.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:56 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:56.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:56 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:56.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:56 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:56.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:56 vm07 ceph-mon[55141]: from='osd.4 [v2:192.168.123.101:6818/3295130738,v1:192.168.123.101:6819/3295130738]' entity='osd.4' 2026-03-06T10:01:56.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:56 vm07 ceph-mon[55141]: from='osd.5 ' entity='osd.5' 2026-03-06T10:01:56.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:56 vm07 ceph-mon[55141]: from='osd.7 [v2:192.168.123.107:6824/1097411483,v1:192.168.123.107:6825/1097411483]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-06T10:01:56.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:56 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:56.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:56 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:56.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:56 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:56.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:56 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:56.452 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:01:56.620 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":24,"num_osds":8,"num_up_osds":4,"osd_up_since":1772787712,"num_in_osds":8,"osd_in_since":1772787694,"num_remapped_pgs":0} 2026-03-06T10:01:57.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:57 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:57.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:57 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:57.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:57 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/1485408030' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:57.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:57 vm07 ceph-mon[55141]: from='osd.7 [v2:192.168.123.107:6824/1097411483,v1:192.168.123.107:6825/1097411483]' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-06T10:01:57.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:57 vm07 ceph-mon[55141]: osd.5 [v2:192.168.123.107:6816/1614548926,v1:192.168.123.107:6817/1614548926] boot 2026-03-06T10:01:57.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:57 vm07 ceph-mon[55141]: osd.4 [v2:192.168.123.101:6818/3295130738,v1:192.168.123.101:6819/3295130738] boot 2026-03-06T10:01:57.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:57 vm07 ceph-mon[55141]: osdmap e25: 8 total, 6 up, 8 in 2026-03-06T10:01:57.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:57 vm07 ceph-mon[55141]: from='osd.7 [v2:192.168.123.107:6824/1097411483,v1:192.168.123.107:6825/1097411483]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-06T10:01:57.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:57 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:57.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:57 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:57.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:57 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:57.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:57 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:57.447 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:57 vm07 ceph-mon[55141]: pgmap v47: 1 pgs: 1 unknown; 0 B data, 106 MiB used, 80 GiB / 80 GiB avail 2026-03-06T10:01:57.620 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd stat -f json 2026-03-06T10:01:57.660 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:57 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:57.660 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:57 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:57.660 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:57 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/1485408030' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:57.660 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:57 vm01 ceph-mon[50931]: from='osd.7 [v2:192.168.123.107:6824/1097411483,v1:192.168.123.107:6825/1097411483]' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-06T10:01:57.660 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:57 vm01 ceph-mon[50931]: osd.5 [v2:192.168.123.107:6816/1614548926,v1:192.168.123.107:6817/1614548926] boot 2026-03-06T10:01:57.660 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:57 vm01 ceph-mon[50931]: osd.4 [v2:192.168.123.101:6818/3295130738,v1:192.168.123.101:6819/3295130738] boot 2026-03-06T10:01:57.660 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:57 vm01 ceph-mon[50931]: osdmap e25: 8 total, 6 up, 8 in 2026-03-06T10:01:57.660 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:57 vm01 ceph-mon[50931]: from='osd.7 [v2:192.168.123.107:6824/1097411483,v1:192.168.123.107:6825/1097411483]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-06T10:01:57.660 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:57 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T10:01:57.660 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:57 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T10:01:57.660 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:57 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:57.660 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:57 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:57.660 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:57 vm01 ceph-mon[50931]: pgmap v47: 1 pgs: 1 unknown; 0 B data, 106 MiB used, 80 GiB / 80 GiB avail 2026-03-06T10:01:58.037 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:01:58.323 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:58 vm01 ceph-mon[50931]: from='osd.6 [v2:192.168.123.101:6826/1332430277,v1:192.168.123.101:6827/1332430277]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-06T10:01:58.323 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:58 vm01 ceph-mon[50931]: from='osd.7 [v2:192.168.123.107:6824/1097411483,v1:192.168.123.107:6825/1097411483]' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-06T10:01:58.323 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:58 vm01 ceph-mon[50931]: from='osd.6 [v2:192.168.123.101:6826/1332430277,v1:192.168.123.101:6827/1332430277]' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-06T10:01:58.323 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:58 vm01 ceph-mon[50931]: osdmap e26: 8 total, 6 up, 8 in 2026-03-06T10:01:58.323 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:58 vm01 ceph-mon[50931]: from='osd.6 [v2:192.168.123.101:6826/1332430277,v1:192.168.123.101:6827/1332430277]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-06T10:01:58.323 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:58 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:58.323 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:58 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:58.323 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:58 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:58.323 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:58 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:58.323 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:58 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:58.323 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:58 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T10:01:58.394 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:01:58.571 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":26,"num_osds":8,"num_up_osds":6,"osd_up_since":1772787716,"num_in_osds":8,"osd_in_since":1772787694,"num_remapped_pgs":0} 2026-03-06T10:01:58.596 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:58 vm07 ceph-mon[55141]: from='osd.6 [v2:192.168.123.101:6826/1332430277,v1:192.168.123.101:6827/1332430277]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-06T10:01:58.596 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:58 vm07 ceph-mon[55141]: from='osd.7 [v2:192.168.123.107:6824/1097411483,v1:192.168.123.107:6825/1097411483]' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-06T10:01:58.596 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:58 vm07 ceph-mon[55141]: from='osd.6 [v2:192.168.123.101:6826/1332430277,v1:192.168.123.101:6827/1332430277]' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-06T10:01:58.596 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:58 vm07 ceph-mon[55141]: osdmap e26: 8 total, 6 up, 8 in 2026-03-06T10:01:58.596 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:58 vm07 ceph-mon[55141]: from='osd.6 [v2:192.168.123.101:6826/1332430277,v1:192.168.123.101:6827/1332430277]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-06T10:01:58.596 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:58 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:58.596 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:58 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:58.597 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:58 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:58.597 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:58 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:58.597 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:58 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:58.597 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:58 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T10:01:59.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:59 vm01 ceph-mon[50931]: purged_snaps scrub starts 2026-03-06T10:01:59.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:59 vm01 ceph-mon[50931]: purged_snaps scrub ok 2026-03-06T10:01:59.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:59 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/1031932986' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:59.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:59 vm01 ceph-mon[50931]: from='osd.7 [v2:192.168.123.107:6824/1097411483,v1:192.168.123.107:6825/1097411483]' entity='osd.7' 2026-03-06T10:01:59.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:59 vm01 ceph-mon[50931]: from='osd.6 [v2:192.168.123.101:6826/1332430277,v1:192.168.123.101:6827/1332430277]' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-06T10:01:59.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:59 vm01 ceph-mon[50931]: osd.7 [v2:192.168.123.107:6824/1097411483,v1:192.168.123.107:6825/1097411483] boot 2026-03-06T10:01:59.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:59 vm01 ceph-mon[50931]: osdmap e27: 8 total, 7 up, 8 in 2026-03-06T10:01:59.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:59 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:59.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:59 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:59.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:59 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:59.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:59 vm01 ceph-mon[50931]: pgmap v50: 1 pgs: 1 active+clean; 449 KiB data, 666 MiB used, 139 GiB / 140 GiB avail 2026-03-06T10:01:59.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:59 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:59.518 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:01:59 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:59.572 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd stat -f json 2026-03-06T10:01:59.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:59 vm07 ceph-mon[55141]: purged_snaps scrub starts 2026-03-06T10:01:59.697 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:59 vm07 ceph-mon[55141]: purged_snaps scrub ok 2026-03-06T10:01:59.697 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:59 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/1031932986' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:01:59.697 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:59 vm07 ceph-mon[55141]: from='osd.7 [v2:192.168.123.107:6824/1097411483,v1:192.168.123.107:6825/1097411483]' entity='osd.7' 2026-03-06T10:01:59.697 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:59 vm07 ceph-mon[55141]: from='osd.6 [v2:192.168.123.101:6826/1332430277,v1:192.168.123.101:6827/1332430277]' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-06T10:01:59.697 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:59 vm07 ceph-mon[55141]: osd.7 [v2:192.168.123.107:6824/1097411483,v1:192.168.123.107:6825/1097411483] boot 2026-03-06T10:01:59.697 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:59 vm07 ceph-mon[55141]: osdmap e27: 8 total, 7 up, 8 in 2026-03-06T10:01:59.697 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:59 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:59.697 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:59 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T10:01:59.697 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:59 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:01:59.697 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:59 vm07 ceph-mon[55141]: pgmap v50: 1 pgs: 1 active+clean; 449 KiB data, 666 MiB used, 139 GiB / 140 GiB avail 2026-03-06T10:01:59.697 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:59 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:01:59.697 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:01:59 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:00.009 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:00.372 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:02:00.550 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":28,"num_osds":8,"num_up_osds":8,"osd_up_since":1772787719,"num_in_osds":8,"osd_in_since":1772787694,"num_remapped_pgs":0} 2026-03-06T10:02:00.551 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd dump --format=json 2026-03-06T10:02:00.579 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:00 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:02:00.579 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:00 vm01 ceph-mon[50931]: osd.6 [v2:192.168.123.101:6826/1332430277,v1:192.168.123.101:6827/1332430277] boot 2026-03-06T10:02:00.579 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:00 vm01 ceph-mon[50931]: osdmap e28: 8 total, 8 up, 8 in 2026-03-06T10:02:00.579 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:00 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:02:00.579 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:00 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:00.579 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:00 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:00.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:00 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:02:00.697 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:00 vm07 ceph-mon[55141]: osd.6 [v2:192.168.123.101:6826/1332430277,v1:192.168.123.101:6827/1332430277] boot 2026-03-06T10:02:00.697 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:00 vm07 ceph-mon[55141]: osdmap e28: 8 total, 8 up, 8 in 2026-03-06T10:02:00.697 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:00 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T10:02:00.697 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:00 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:00.697 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:00 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:00.951 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:01.321 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:02:01.321 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":28,"fsid":"a3ab6672-193a-11f1-b81f-a119763c7190","created":"2026-03-06T08:59:13.186693+0000","modified":"2026-03-06T09:01:59.630967+0000","last_up_change":"2026-03-06T09:01:59.630967+0000","last_in_change":"2026-03-06T09:01:34.704237+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":13,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-06T09:01:52.856926+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"24","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":7.8899998664855957,"score_stable":7.8899998664855957,"optimal_score":0.37999999523162842,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"5419104f-76e4-4e2f-9035-ff0ac04e9fed","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":16,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6800","nonce":4015024690},{"type":"v1","addr":"192.168.123.107:6801","nonce":4015024690}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6802","nonce":4015024690},{"type":"v1","addr":"192.168.123.107:6803","nonce":4015024690}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6806","nonce":4015024690},{"type":"v1","addr":"192.168.123.107:6807","nonce":4015024690}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6804","nonce":4015024690},{"type":"v1","addr":"192.168.123.107:6805","nonce":4015024690}]},"public_addr":"192.168.123.107:6801/4015024690","cluster_addr":"192.168.123.107:6803/4015024690","heartbeat_back_addr":"192.168.123.107:6807/4015024690","heartbeat_front_addr":"192.168.123.107:6805/4015024690","state":["exists","up"]},{"osd":1,"uuid":"f639e6c1-958d-4147-b8c2-bf8bb5af0c1e","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":22,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6802","nonce":3951729305},{"type":"v1","addr":"192.168.123.101:6803","nonce":3951729305}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6804","nonce":3951729305},{"type":"v1","addr":"192.168.123.101:6805","nonce":3951729305}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6808","nonce":3951729305},{"type":"v1","addr":"192.168.123.101:6809","nonce":3951729305}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6806","nonce":3951729305},{"type":"v1","addr":"192.168.123.101:6807","nonce":3951729305}]},"public_addr":"192.168.123.101:6803/3951729305","cluster_addr":"192.168.123.101:6805/3951729305","heartbeat_back_addr":"192.168.123.101:6809/3951729305","heartbeat_front_addr":"192.168.123.101:6807/3951729305","state":["exists","up"]},{"osd":2,"uuid":"d645274a-f25c-4360-be0f-460223194536","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":21,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6810","nonce":970532629},{"type":"v1","addr":"192.168.123.101:6811","nonce":970532629}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6812","nonce":970532629},{"type":"v1","addr":"192.168.123.101:6813","nonce":970532629}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6816","nonce":970532629},{"type":"v1","addr":"192.168.123.101:6817","nonce":970532629}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6814","nonce":970532629},{"type":"v1","addr":"192.168.123.101:6815","nonce":970532629}]},"public_addr":"192.168.123.101:6811/970532629","cluster_addr":"192.168.123.101:6813/970532629","heartbeat_back_addr":"192.168.123.101:6817/970532629","heartbeat_front_addr":"192.168.123.101:6815/970532629","state":["exists","up"]},{"osd":3,"uuid":"75751174-b0ab-4614-a691-06b4cd42e273","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":20,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6808","nonce":206187849},{"type":"v1","addr":"192.168.123.107:6809","nonce":206187849}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6810","nonce":206187849},{"type":"v1","addr":"192.168.123.107:6811","nonce":206187849}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6814","nonce":206187849},{"type":"v1","addr":"192.168.123.107:6815","nonce":206187849}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6812","nonce":206187849},{"type":"v1","addr":"192.168.123.107:6813","nonce":206187849}]},"public_addr":"192.168.123.107:6809/206187849","cluster_addr":"192.168.123.107:6811/206187849","heartbeat_back_addr":"192.168.123.107:6815/206187849","heartbeat_front_addr":"192.168.123.107:6813/206187849","state":["exists","up"]},{"osd":4,"uuid":"84dec1a3-932d-4894-b752-0323fbc97b37","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":25,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6818","nonce":3295130738},{"type":"v1","addr":"192.168.123.101:6819","nonce":3295130738}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6820","nonce":3295130738},{"type":"v1","addr":"192.168.123.101:6821","nonce":3295130738}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6824","nonce":3295130738},{"type":"v1","addr":"192.168.123.101:6825","nonce":3295130738}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6822","nonce":3295130738},{"type":"v1","addr":"192.168.123.101:6823","nonce":3295130738}]},"public_addr":"192.168.123.101:6819/3295130738","cluster_addr":"192.168.123.101:6821/3295130738","heartbeat_back_addr":"192.168.123.101:6825/3295130738","heartbeat_front_addr":"192.168.123.101:6823/3295130738","state":["exists","up"]},{"osd":5,"uuid":"036ff726-417f-464a-b30b-e5d7e55a059f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":25,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6816","nonce":1614548926},{"type":"v1","addr":"192.168.123.107:6817","nonce":1614548926}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6818","nonce":1614548926},{"type":"v1","addr":"192.168.123.107:6819","nonce":1614548926}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6822","nonce":1614548926},{"type":"v1","addr":"192.168.123.107:6823","nonce":1614548926}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6820","nonce":1614548926},{"type":"v1","addr":"192.168.123.107:6821","nonce":1614548926}]},"public_addr":"192.168.123.107:6817/1614548926","cluster_addr":"192.168.123.107:6819/1614548926","heartbeat_back_addr":"192.168.123.107:6823/1614548926","heartbeat_front_addr":"192.168.123.107:6821/1614548926","state":["exists","up"]},{"osd":6,"uuid":"1705c5f8-0f21-4e35-a9df-ef5ae49cc4c3","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":28,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6826","nonce":1332430277},{"type":"v1","addr":"192.168.123.101:6827","nonce":1332430277}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6828","nonce":1332430277},{"type":"v1","addr":"192.168.123.101:6829","nonce":1332430277}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6832","nonce":1332430277},{"type":"v1","addr":"192.168.123.101:6833","nonce":1332430277}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6830","nonce":1332430277},{"type":"v1","addr":"192.168.123.101:6831","nonce":1332430277}]},"public_addr":"192.168.123.101:6827/1332430277","cluster_addr":"192.168.123.101:6829/1332430277","heartbeat_back_addr":"192.168.123.101:6833/1332430277","heartbeat_front_addr":"192.168.123.101:6831/1332430277","state":["exists","up"]},{"osd":7,"uuid":"e96c75da-b3cf-40d0-930b-2665d7b20d97","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":27,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6824","nonce":1097411483},{"type":"v1","addr":"192.168.123.107:6825","nonce":1097411483}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6826","nonce":1097411483},{"type":"v1","addr":"192.168.123.107:6827","nonce":1097411483}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6830","nonce":1097411483},{"type":"v1","addr":"192.168.123.107:6831","nonce":1097411483}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6828","nonce":1097411483},{"type":"v1","addr":"192.168.123.107:6829","nonce":1097411483}]},"public_addr":"192.168.123.107:6825/1097411483","cluster_addr":"192.168.123.107:6827/1097411483","heartbeat_back_addr":"192.168.123.107:6831/1097411483","heartbeat_front_addr":"192.168.123.107:6829/1097411483","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T09:01:45.773734+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T09:01:46.206756+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T09:01:50.979420+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T09:01:49.319216+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T09:01:53.931329+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T09:01:53.159214+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T09:01:56.693762+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.101:0/3083305098":"2026-03-07T09:00:48.806129+0000","192.168.123.101:0/1779951470":"2026-03-07T09:00:48.806129+0000","192.168.123.101:6800/1918328609":"2026-03-07T08:59:42.760014+0000","192.168.123.101:6801/1918328609":"2026-03-07T08:59:42.760014+0000","192.168.123.101:0/1767720239":"2026-03-07T08:59:42.760014+0000","192.168.123.101:6801/1478391393":"2026-03-07T09:00:48.806129+0000","192.168.123.101:0/3577657292":"2026-03-07T09:00:06.093174+0000","192.168.123.101:0/4214858901":"2026-03-07T08:59:42.760014+0000","192.168.123.101:0/3607041991":"2026-03-07T08:59:42.760014+0000","192.168.123.101:6800/600350983":"2026-03-07T09:00:06.093174+0000","192.168.123.101:0/1404783458":"2026-03-07T09:00:06.093174+0000","192.168.123.101:6801/600350983":"2026-03-07T09:00:06.093174+0000","192.168.123.101:0/787451282":"2026-03-07T09:00:06.093174+0000","192.168.123.101:0/623415636":"2026-03-07T09:00:48.806129+0000","192.168.123.101:6800/1478391393":"2026-03-07T09:00:48.806129+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-06T10:02:01.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:01 vm07 ceph-mon[55141]: purged_snaps scrub starts 2026-03-06T10:02:01.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:01 vm07 ceph-mon[55141]: purged_snaps scrub ok 2026-03-06T10:02:01.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:01 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/2538397460' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:02:01.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:01 vm07 ceph-mon[55141]: pgmap v52: 1 pgs: 1 active+clean; 449 KiB data, 693 MiB used, 159 GiB / 160 GiB avail 2026-03-06T10:02:01.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:01 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:01.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:01 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:01.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:01 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config rm", "who": "osd/host:vm07", "name": "osd_memory_target"}]: dispatch 2026-03-06T10:02:01.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:01 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/2220035565' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T10:02:01.500 INFO:tasks.cephadm.ceph_manager.ceph:[{'pool': 1, 'pool_name': '.mgr', 'create_time': '2026-03-06T09:01:52.856926+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'is_stretch_pool': False, 'object_hash': 2, 'pg_autoscale_mode': 'off', 'pg_num': 1, 'pg_placement_num': 1, 'pg_placement_num_target': 1, 'pg_num_target': 1, 'pg_num_pending': 1, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': "0'0", 'target_version': "0'0"}, 'last_change': '24', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {'pg_num_max': 32, 'pg_num_min': 1}, 'application_metadata': {'mgr': {}}, 'read_balance': {'score_type': 'Fair distribution', 'score_acting': 7.889999866485596, 'score_stable': 7.889999866485596, 'optimal_score': 0.3799999952316284, 'raw_score_acting': 3, 'raw_score_stable': 3, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}}] 2026-03-06T10:02:01.500 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd pool get .mgr pg_num 2026-03-06T10:02:01.572 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:01 vm01 ceph-mon[50931]: purged_snaps scrub starts 2026-03-06T10:02:01.572 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:01 vm01 ceph-mon[50931]: purged_snaps scrub ok 2026-03-06T10:02:01.572 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:01 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/2538397460' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T10:02:01.572 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:01 vm01 ceph-mon[50931]: pgmap v52: 1 pgs: 1 active+clean; 449 KiB data, 693 MiB used, 159 GiB / 160 GiB avail 2026-03-06T10:02:01.572 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:01 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:01.572 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:01 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:01.572 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:01 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config rm", "who": "osd/host:vm07", "name": "osd_memory_target"}]: dispatch 2026-03-06T10:02:01.572 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:01 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/2220035565' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T10:02:01.929 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:02.367 INFO:teuthology.orchestra.run.vm01.stdout:pg_num: 1 2026-03-06T10:02:02.494 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:02 vm01 ceph-mon[50931]: Detected new or changed devices on vm07 2026-03-06T10:02:02.494 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:02 vm01 ceph-mon[50931]: osdmap e29: 8 total, 8 up, 8 in 2026-03-06T10:02:02.494 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:02 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:02.494 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:02 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:02.494 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:02 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-06T10:02:02.494 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:02 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:02:02.494 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:02 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T10:02:02.494 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:02 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:02.494 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:02 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T10:02:02.526 INFO:tasks.cephadm:Setting up client nodes... 2026-03-06T10:02:02.527 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph auth get-or-create client.0 mon 'allow *' osd 'allow *' mds 'allow *' mgr 'allow *' 2026-03-06T10:02:02.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:02 vm07 ceph-mon[55141]: Detected new or changed devices on vm07 2026-03-06T10:02:02.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:02 vm07 ceph-mon[55141]: osdmap e29: 8 total, 8 up, 8 in 2026-03-06T10:02:02.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:02 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:02.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:02 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:02.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:02 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-06T10:02:02.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:02 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:02:02.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:02 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T10:02:02.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:02 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:02.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:02 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T10:02:02.891 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:03.254 INFO:teuthology.orchestra.run.vm01.stdout:[client.0] 2026-03-06T10:02:03.254 INFO:teuthology.orchestra.run.vm01.stdout: key = AQALmKpp0eHqDhAAMad3DHUfhiuk4Ivf1UesJA== 2026-03-06T10:02:03.424 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T10:02:03.424 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/ceph/ceph.client.0.keyring 2026-03-06T10:02:03.424 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod 0644 /etc/ceph/ceph.client.0.keyring 2026-03-06T10:02:03.458 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph auth get-or-create client.1 mon 'allow *' osd 'allow *' mds 'allow *' mgr 'allow *' 2026-03-06T10:02:03.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:03 vm01 ceph-mon[50931]: Detected new or changed devices on vm01 2026-03-06T10:02:03.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:03 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/2663163816' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-06T10:02:03.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:03 vm01 ceph-mon[50931]: pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 293 MiB used, 160 GiB / 160 GiB avail 2026-03-06T10:02:03.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:03 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/3676854254' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-06T10:02:03.517 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:03 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/3676854254' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-06T10:02:03.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:03 vm07 ceph-mon[55141]: Detected new or changed devices on vm01 2026-03-06T10:02:03.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:03 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/2663163816' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-06T10:02:03.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:03 vm07 ceph-mon[55141]: pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 293 MiB used, 160 GiB / 160 GiB avail 2026-03-06T10:02:03.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:03 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/3676854254' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-06T10:02:03.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:03 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/3676854254' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-06T10:02:03.770 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm07/config 2026-03-06T10:02:04.136 INFO:teuthology.orchestra.run.vm07.stdout:[client.1] 2026-03-06T10:02:04.136 INFO:teuthology.orchestra.run.vm07.stdout: key = AQAMmKppZ37JBxAAy0zorzxd4dm/jpGfJNsfhQ== 2026-03-06T10:02:04.286 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-06T10:02:04.286 DEBUG:teuthology.orchestra.run.vm07:> sudo dd of=/etc/ceph/ceph.client.1.keyring 2026-03-06T10:02:04.286 DEBUG:teuthology.orchestra.run.vm07:> sudo chmod 0644 /etc/ceph/ceph.client.1.keyring 2026-03-06T10:02:04.317 INFO:tasks.ceph:Waiting until ceph daemons up and pgs clean... 2026-03-06T10:02:04.317 INFO:tasks.cephadm.ceph_manager.ceph:waiting for mgr available 2026-03-06T10:02:04.317 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph mgr dump --format=json 2026-03-06T10:02:04.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:04 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T10:02:04.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:04 vm07 ceph-mon[55141]: from='client.? 192.168.123.107:0/2921924467' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-06T10:02:04.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:04 vm07 ceph-mon[55141]: from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-06T10:02:04.446 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:04 vm07 ceph-mon[55141]: from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-06T10:02:04.656 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:04.680 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:04 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T10:02:04.680 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:04 vm01 ceph-mon[50931]: from='client.? 192.168.123.107:0/2921924467' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-06T10:02:04.680 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:04 vm01 ceph-mon[50931]: from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-06T10:02:04.680 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:04 vm01 ceph-mon[50931]: from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-06T10:02:05.030 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:02:05.206 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":19,"flags":0,"active_gid":14219,"active_name":"vm01.ifwqbh","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6800","nonce":2567257091},{"type":"v1","addr":"192.168.123.101:6801","nonce":2567257091}]},"active_addr":"192.168.123.101:6801/2567257091","active_change":"2026-03-06T09:00:48.806447+0000","active_mgr_features":4540701547738038271,"available":true,"standbys":[{"gid":14250,"name":"vm07.myglqt","mgr_features":4540701547738038271,"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}]}],"modules":["cephadm","dashboard","iostat","nfs","prometheus","restful"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{"dashboard":"https://192.168.123.101:8443/","prometheus":"http://192.168.123.101:9283/"},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"reef":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"squid":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"]},"force_disabled_modules":{},"last_failure_osd_epoch":5,"active_clients":[{"name":"devicehealth","addrvec":[{"type":"v2","addr":"192.168.123.101:0","nonce":3823746628}]},{"name":"libcephsqlite","addrvec":[{"type":"v2","addr":"192.168.123.101:0","nonce":3741748152}]},{"name":"rbd_support","addrvec":[{"type":"v2","addr":"192.168.123.101:0","nonce":1121347101}]},{"name":"volumes","addrvec":[{"type":"v2","addr":"192.168.123.101:0","nonce":2487402294}]}]} 2026-03-06T10:02:05.208 INFO:tasks.cephadm.ceph_manager.ceph:mgr available! 2026-03-06T10:02:05.208 INFO:tasks.cephadm.ceph_manager.ceph:waiting for all up 2026-03-06T10:02:05.208 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd dump --format=json 2026-03-06T10:02:05.353 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:05 vm01 ceph-mon[50931]: pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-06T10:02:05.353 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:05 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/560179064' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-06T10:02:05.539 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:05.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:05 vm07 ceph-mon[55141]: pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-06T10:02:05.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:05 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/560179064' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-06T10:02:05.866 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:02:05.866 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":29,"fsid":"a3ab6672-193a-11f1-b81f-a119763c7190","created":"2026-03-06T08:59:13.186693+0000","modified":"2026-03-06T09:02:01.329608+0000","last_up_change":"2026-03-06T09:01:59.630967+0000","last_in_change":"2026-03-06T09:01:34.704237+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":13,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-06T09:01:52.856926+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"24","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":7.8899998664855957,"score_stable":7.8899998664855957,"optimal_score":0.37999999523162842,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"5419104f-76e4-4e2f-9035-ff0ac04e9fed","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":16,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6800","nonce":4015024690},{"type":"v1","addr":"192.168.123.107:6801","nonce":4015024690}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6802","nonce":4015024690},{"type":"v1","addr":"192.168.123.107:6803","nonce":4015024690}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6806","nonce":4015024690},{"type":"v1","addr":"192.168.123.107:6807","nonce":4015024690}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6804","nonce":4015024690},{"type":"v1","addr":"192.168.123.107:6805","nonce":4015024690}]},"public_addr":"192.168.123.107:6801/4015024690","cluster_addr":"192.168.123.107:6803/4015024690","heartbeat_back_addr":"192.168.123.107:6807/4015024690","heartbeat_front_addr":"192.168.123.107:6805/4015024690","state":["exists","up"]},{"osd":1,"uuid":"f639e6c1-958d-4147-b8c2-bf8bb5af0c1e","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":22,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6802","nonce":3951729305},{"type":"v1","addr":"192.168.123.101:6803","nonce":3951729305}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6804","nonce":3951729305},{"type":"v1","addr":"192.168.123.101:6805","nonce":3951729305}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6808","nonce":3951729305},{"type":"v1","addr":"192.168.123.101:6809","nonce":3951729305}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6806","nonce":3951729305},{"type":"v1","addr":"192.168.123.101:6807","nonce":3951729305}]},"public_addr":"192.168.123.101:6803/3951729305","cluster_addr":"192.168.123.101:6805/3951729305","heartbeat_back_addr":"192.168.123.101:6809/3951729305","heartbeat_front_addr":"192.168.123.101:6807/3951729305","state":["exists","up"]},{"osd":2,"uuid":"d645274a-f25c-4360-be0f-460223194536","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":21,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6810","nonce":970532629},{"type":"v1","addr":"192.168.123.101:6811","nonce":970532629}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6812","nonce":970532629},{"type":"v1","addr":"192.168.123.101:6813","nonce":970532629}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6816","nonce":970532629},{"type":"v1","addr":"192.168.123.101:6817","nonce":970532629}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6814","nonce":970532629},{"type":"v1","addr":"192.168.123.101:6815","nonce":970532629}]},"public_addr":"192.168.123.101:6811/970532629","cluster_addr":"192.168.123.101:6813/970532629","heartbeat_back_addr":"192.168.123.101:6817/970532629","heartbeat_front_addr":"192.168.123.101:6815/970532629","state":["exists","up"]},{"osd":3,"uuid":"75751174-b0ab-4614-a691-06b4cd42e273","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":20,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6808","nonce":206187849},{"type":"v1","addr":"192.168.123.107:6809","nonce":206187849}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6810","nonce":206187849},{"type":"v1","addr":"192.168.123.107:6811","nonce":206187849}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6814","nonce":206187849},{"type":"v1","addr":"192.168.123.107:6815","nonce":206187849}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6812","nonce":206187849},{"type":"v1","addr":"192.168.123.107:6813","nonce":206187849}]},"public_addr":"192.168.123.107:6809/206187849","cluster_addr":"192.168.123.107:6811/206187849","heartbeat_back_addr":"192.168.123.107:6815/206187849","heartbeat_front_addr":"192.168.123.107:6813/206187849","state":["exists","up"]},{"osd":4,"uuid":"84dec1a3-932d-4894-b752-0323fbc97b37","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":25,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6818","nonce":3295130738},{"type":"v1","addr":"192.168.123.101:6819","nonce":3295130738}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6820","nonce":3295130738},{"type":"v1","addr":"192.168.123.101:6821","nonce":3295130738}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6824","nonce":3295130738},{"type":"v1","addr":"192.168.123.101:6825","nonce":3295130738}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6822","nonce":3295130738},{"type":"v1","addr":"192.168.123.101:6823","nonce":3295130738}]},"public_addr":"192.168.123.101:6819/3295130738","cluster_addr":"192.168.123.101:6821/3295130738","heartbeat_back_addr":"192.168.123.101:6825/3295130738","heartbeat_front_addr":"192.168.123.101:6823/3295130738","state":["exists","up"]},{"osd":5,"uuid":"036ff726-417f-464a-b30b-e5d7e55a059f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":25,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6816","nonce":1614548926},{"type":"v1","addr":"192.168.123.107:6817","nonce":1614548926}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6818","nonce":1614548926},{"type":"v1","addr":"192.168.123.107:6819","nonce":1614548926}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6822","nonce":1614548926},{"type":"v1","addr":"192.168.123.107:6823","nonce":1614548926}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6820","nonce":1614548926},{"type":"v1","addr":"192.168.123.107:6821","nonce":1614548926}]},"public_addr":"192.168.123.107:6817/1614548926","cluster_addr":"192.168.123.107:6819/1614548926","heartbeat_back_addr":"192.168.123.107:6823/1614548926","heartbeat_front_addr":"192.168.123.107:6821/1614548926","state":["exists","up"]},{"osd":6,"uuid":"1705c5f8-0f21-4e35-a9df-ef5ae49cc4c3","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":28,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6826","nonce":1332430277},{"type":"v1","addr":"192.168.123.101:6827","nonce":1332430277}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6828","nonce":1332430277},{"type":"v1","addr":"192.168.123.101:6829","nonce":1332430277}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6832","nonce":1332430277},{"type":"v1","addr":"192.168.123.101:6833","nonce":1332430277}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6830","nonce":1332430277},{"type":"v1","addr":"192.168.123.101:6831","nonce":1332430277}]},"public_addr":"192.168.123.101:6827/1332430277","cluster_addr":"192.168.123.101:6829/1332430277","heartbeat_back_addr":"192.168.123.101:6833/1332430277","heartbeat_front_addr":"192.168.123.101:6831/1332430277","state":["exists","up"]},{"osd":7,"uuid":"e96c75da-b3cf-40d0-930b-2665d7b20d97","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":27,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6824","nonce":1097411483},{"type":"v1","addr":"192.168.123.107:6825","nonce":1097411483}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6826","nonce":1097411483},{"type":"v1","addr":"192.168.123.107:6827","nonce":1097411483}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6830","nonce":1097411483},{"type":"v1","addr":"192.168.123.107:6831","nonce":1097411483}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6828","nonce":1097411483},{"type":"v1","addr":"192.168.123.107:6829","nonce":1097411483}]},"public_addr":"192.168.123.107:6825/1097411483","cluster_addr":"192.168.123.107:6827/1097411483","heartbeat_back_addr":"192.168.123.107:6831/1097411483","heartbeat_front_addr":"192.168.123.107:6829/1097411483","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T09:01:45.773734+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T09:01:46.206756+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T09:01:50.979420+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T09:01:49.319216+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T09:01:53.931329+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T09:01:53.159214+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T09:01:58.511749+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T09:01:56.693762+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.101:0/3083305098":"2026-03-07T09:00:48.806129+0000","192.168.123.101:0/1779951470":"2026-03-07T09:00:48.806129+0000","192.168.123.101:6800/1918328609":"2026-03-07T08:59:42.760014+0000","192.168.123.101:6801/1918328609":"2026-03-07T08:59:42.760014+0000","192.168.123.101:0/1767720239":"2026-03-07T08:59:42.760014+0000","192.168.123.101:6801/1478391393":"2026-03-07T09:00:48.806129+0000","192.168.123.101:0/3577657292":"2026-03-07T09:00:06.093174+0000","192.168.123.101:0/4214858901":"2026-03-07T08:59:42.760014+0000","192.168.123.101:0/3607041991":"2026-03-07T08:59:42.760014+0000","192.168.123.101:6800/600350983":"2026-03-07T09:00:06.093174+0000","192.168.123.101:0/1404783458":"2026-03-07T09:00:06.093174+0000","192.168.123.101:6801/600350983":"2026-03-07T09:00:06.093174+0000","192.168.123.101:0/787451282":"2026-03-07T09:00:06.093174+0000","192.168.123.101:0/623415636":"2026-03-07T09:00:48.806129+0000","192.168.123.101:6800/1478391393":"2026-03-07T09:00:48.806129+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-06T10:02:06.041 INFO:tasks.cephadm.ceph_manager.ceph:all up! 2026-03-06T10:02:06.041 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd dump --format=json 2026-03-06T10:02:06.365 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:06.691 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:06 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/3601795986' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T10:02:06.691 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:02:06.691 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":29,"fsid":"a3ab6672-193a-11f1-b81f-a119763c7190","created":"2026-03-06T08:59:13.186693+0000","modified":"2026-03-06T09:02:01.329608+0000","last_up_change":"2026-03-06T09:01:59.630967+0000","last_in_change":"2026-03-06T09:01:34.704237+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":13,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-06T09:01:52.856926+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"24","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":7.8899998664855957,"score_stable":7.8899998664855957,"optimal_score":0.37999999523162842,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"5419104f-76e4-4e2f-9035-ff0ac04e9fed","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":16,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6800","nonce":4015024690},{"type":"v1","addr":"192.168.123.107:6801","nonce":4015024690}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6802","nonce":4015024690},{"type":"v1","addr":"192.168.123.107:6803","nonce":4015024690}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6806","nonce":4015024690},{"type":"v1","addr":"192.168.123.107:6807","nonce":4015024690}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6804","nonce":4015024690},{"type":"v1","addr":"192.168.123.107:6805","nonce":4015024690}]},"public_addr":"192.168.123.107:6801/4015024690","cluster_addr":"192.168.123.107:6803/4015024690","heartbeat_back_addr":"192.168.123.107:6807/4015024690","heartbeat_front_addr":"192.168.123.107:6805/4015024690","state":["exists","up"]},{"osd":1,"uuid":"f639e6c1-958d-4147-b8c2-bf8bb5af0c1e","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":22,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6802","nonce":3951729305},{"type":"v1","addr":"192.168.123.101:6803","nonce":3951729305}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6804","nonce":3951729305},{"type":"v1","addr":"192.168.123.101:6805","nonce":3951729305}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6808","nonce":3951729305},{"type":"v1","addr":"192.168.123.101:6809","nonce":3951729305}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6806","nonce":3951729305},{"type":"v1","addr":"192.168.123.101:6807","nonce":3951729305}]},"public_addr":"192.168.123.101:6803/3951729305","cluster_addr":"192.168.123.101:6805/3951729305","heartbeat_back_addr":"192.168.123.101:6809/3951729305","heartbeat_front_addr":"192.168.123.101:6807/3951729305","state":["exists","up"]},{"osd":2,"uuid":"d645274a-f25c-4360-be0f-460223194536","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":21,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6810","nonce":970532629},{"type":"v1","addr":"192.168.123.101:6811","nonce":970532629}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6812","nonce":970532629},{"type":"v1","addr":"192.168.123.101:6813","nonce":970532629}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6816","nonce":970532629},{"type":"v1","addr":"192.168.123.101:6817","nonce":970532629}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6814","nonce":970532629},{"type":"v1","addr":"192.168.123.101:6815","nonce":970532629}]},"public_addr":"192.168.123.101:6811/970532629","cluster_addr":"192.168.123.101:6813/970532629","heartbeat_back_addr":"192.168.123.101:6817/970532629","heartbeat_front_addr":"192.168.123.101:6815/970532629","state":["exists","up"]},{"osd":3,"uuid":"75751174-b0ab-4614-a691-06b4cd42e273","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":20,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6808","nonce":206187849},{"type":"v1","addr":"192.168.123.107:6809","nonce":206187849}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6810","nonce":206187849},{"type":"v1","addr":"192.168.123.107:6811","nonce":206187849}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6814","nonce":206187849},{"type":"v1","addr":"192.168.123.107:6815","nonce":206187849}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6812","nonce":206187849},{"type":"v1","addr":"192.168.123.107:6813","nonce":206187849}]},"public_addr":"192.168.123.107:6809/206187849","cluster_addr":"192.168.123.107:6811/206187849","heartbeat_back_addr":"192.168.123.107:6815/206187849","heartbeat_front_addr":"192.168.123.107:6813/206187849","state":["exists","up"]},{"osd":4,"uuid":"84dec1a3-932d-4894-b752-0323fbc97b37","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":25,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6818","nonce":3295130738},{"type":"v1","addr":"192.168.123.101:6819","nonce":3295130738}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6820","nonce":3295130738},{"type":"v1","addr":"192.168.123.101:6821","nonce":3295130738}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6824","nonce":3295130738},{"type":"v1","addr":"192.168.123.101:6825","nonce":3295130738}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6822","nonce":3295130738},{"type":"v1","addr":"192.168.123.101:6823","nonce":3295130738}]},"public_addr":"192.168.123.101:6819/3295130738","cluster_addr":"192.168.123.101:6821/3295130738","heartbeat_back_addr":"192.168.123.101:6825/3295130738","heartbeat_front_addr":"192.168.123.101:6823/3295130738","state":["exists","up"]},{"osd":5,"uuid":"036ff726-417f-464a-b30b-e5d7e55a059f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":25,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6816","nonce":1614548926},{"type":"v1","addr":"192.168.123.107:6817","nonce":1614548926}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6818","nonce":1614548926},{"type":"v1","addr":"192.168.123.107:6819","nonce":1614548926}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6822","nonce":1614548926},{"type":"v1","addr":"192.168.123.107:6823","nonce":1614548926}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6820","nonce":1614548926},{"type":"v1","addr":"192.168.123.107:6821","nonce":1614548926}]},"public_addr":"192.168.123.107:6817/1614548926","cluster_addr":"192.168.123.107:6819/1614548926","heartbeat_back_addr":"192.168.123.107:6823/1614548926","heartbeat_front_addr":"192.168.123.107:6821/1614548926","state":["exists","up"]},{"osd":6,"uuid":"1705c5f8-0f21-4e35-a9df-ef5ae49cc4c3","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":28,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6826","nonce":1332430277},{"type":"v1","addr":"192.168.123.101:6827","nonce":1332430277}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6828","nonce":1332430277},{"type":"v1","addr":"192.168.123.101:6829","nonce":1332430277}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6832","nonce":1332430277},{"type":"v1","addr":"192.168.123.101:6833","nonce":1332430277}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6830","nonce":1332430277},{"type":"v1","addr":"192.168.123.101:6831","nonce":1332430277}]},"public_addr":"192.168.123.101:6827/1332430277","cluster_addr":"192.168.123.101:6829/1332430277","heartbeat_back_addr":"192.168.123.101:6833/1332430277","heartbeat_front_addr":"192.168.123.101:6831/1332430277","state":["exists","up"]},{"osd":7,"uuid":"e96c75da-b3cf-40d0-930b-2665d7b20d97","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":27,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6824","nonce":1097411483},{"type":"v1","addr":"192.168.123.107:6825","nonce":1097411483}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6826","nonce":1097411483},{"type":"v1","addr":"192.168.123.107:6827","nonce":1097411483}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6830","nonce":1097411483},{"type":"v1","addr":"192.168.123.107:6831","nonce":1097411483}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6828","nonce":1097411483},{"type":"v1","addr":"192.168.123.107:6829","nonce":1097411483}]},"public_addr":"192.168.123.107:6825/1097411483","cluster_addr":"192.168.123.107:6827/1097411483","heartbeat_back_addr":"192.168.123.107:6831/1097411483","heartbeat_front_addr":"192.168.123.107:6829/1097411483","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T09:01:45.773734+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T09:01:46.206756+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T09:01:50.979420+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T09:01:49.319216+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T09:01:53.931329+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T09:01:53.159214+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T09:01:58.511749+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T09:01:56.693762+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.101:0/3083305098":"2026-03-07T09:00:48.806129+0000","192.168.123.101:0/1779951470":"2026-03-07T09:00:48.806129+0000","192.168.123.101:6800/1918328609":"2026-03-07T08:59:42.760014+0000","192.168.123.101:6801/1918328609":"2026-03-07T08:59:42.760014+0000","192.168.123.101:0/1767720239":"2026-03-07T08:59:42.760014+0000","192.168.123.101:6801/1478391393":"2026-03-07T09:00:48.806129+0000","192.168.123.101:0/3577657292":"2026-03-07T09:00:06.093174+0000","192.168.123.101:0/4214858901":"2026-03-07T08:59:42.760014+0000","192.168.123.101:0/3607041991":"2026-03-07T08:59:42.760014+0000","192.168.123.101:6800/600350983":"2026-03-07T09:00:06.093174+0000","192.168.123.101:0/1404783458":"2026-03-07T09:00:06.093174+0000","192.168.123.101:6801/600350983":"2026-03-07T09:00:06.093174+0000","192.168.123.101:0/787451282":"2026-03-07T09:00:06.093174+0000","192.168.123.101:0/623415636":"2026-03-07T09:00:48.806129+0000","192.168.123.101:6800/1478391393":"2026-03-07T09:00:48.806129+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-06T10:02:06.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:06 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/3601795986' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T10:02:06.845 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph tell osd.0 flush_pg_stats 2026-03-06T10:02:06.846 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph tell osd.1 flush_pg_stats 2026-03-06T10:02:06.846 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph tell osd.2 flush_pg_stats 2026-03-06T10:02:06.846 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph tell osd.3 flush_pg_stats 2026-03-06T10:02:06.846 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph tell osd.4 flush_pg_stats 2026-03-06T10:02:06.846 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph tell osd.5 flush_pg_stats 2026-03-06T10:02:06.846 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph tell osd.6 flush_pg_stats 2026-03-06T10:02:06.846 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph tell osd.7 flush_pg_stats 2026-03-06T10:02:07.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:07 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/2410933566' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T10:02:07.696 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:07 vm07 ceph-mon[55141]: pgmap v56: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-06T10:02:07.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:07 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/2410933566' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T10:02:07.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:07 vm01 ceph-mon[50931]: pgmap v56: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-06T10:02:08.272 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:08.281 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:08.307 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:08.310 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:08.324 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:08.514 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:08.530 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:08.532 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:09.588 INFO:teuthology.orchestra.run.vm01.stdout:73014444038 2026-03-06T10:02:09.589 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd last-stat-seq osd.1 2026-03-06T10:02:09.675 INFO:teuthology.orchestra.run.vm01.stdout:68719476742 2026-03-06T10:02:09.675 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd last-stat-seq osd.0 2026-03-06T10:02:09.748 INFO:teuthology.orchestra.run.vm01.stdout:90194313221 2026-03-06T10:02:09.748 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd last-stat-seq osd.2 2026-03-06T10:02:09.868 INFO:teuthology.orchestra.run.vm01.stdout:85899345925 2026-03-06T10:02:09.868 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd last-stat-seq osd.3 2026-03-06T10:02:09.925 INFO:teuthology.orchestra.run.vm01.stdout:107374182404 2026-03-06T10:02:09.925 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd last-stat-seq osd.4 2026-03-06T10:02:10.015 INFO:teuthology.orchestra.run.vm01.stdout:120259084291 2026-03-06T10:02:10.015 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd last-stat-seq osd.6 2026-03-06T10:02:10.043 INFO:teuthology.orchestra.run.vm01.stdout:115964116996 2026-03-06T10:02:10.043 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd last-stat-seq osd.7 2026-03-06T10:02:10.049 INFO:teuthology.orchestra.run.vm01.stdout:107374182404 2026-03-06T10:02:10.049 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph osd last-stat-seq osd.5 2026-03-06T10:02:10.188 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:09 vm01 ceph-mon[50931]: pgmap v57: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-06T10:02:10.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:09 vm07 ceph-mon[55141]: pgmap v57: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-06T10:02:11.087 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:11.129 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:11.425 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:11.607 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:11.787 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:11.790 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:11.955 INFO:teuthology.orchestra.run.vm01.stdout:85899345925 2026-03-06T10:02:12.086 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:11 vm01 ceph-mon[50931]: pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-06T10:02:12.087 INFO:teuthology.orchestra.run.vm01.stdout:73014444038 2026-03-06T10:02:12.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:11 vm07 ceph-mon[55141]: pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-06T10:02:12.233 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:12.255 INFO:tasks.cephadm.ceph_manager.ceph:need seq 85899345925 got 85899345925 for osd.3 2026-03-06T10:02:12.255 DEBUG:teuthology.parallel:result is None 2026-03-06T10:02:12.301 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:12.381 INFO:tasks.cephadm.ceph_manager.ceph:need seq 73014444038 got 73014444038 for osd.1 2026-03-06T10:02:12.381 DEBUG:teuthology.parallel:result is None 2026-03-06T10:02:12.699 INFO:teuthology.orchestra.run.vm01.stdout:120259084292 2026-03-06T10:02:12.829 INFO:teuthology.orchestra.run.vm01.stdout:90194313221 2026-03-06T10:02:12.887 INFO:tasks.cephadm.ceph_manager.ceph:need seq 120259084291 got 120259084292 for osd.6 2026-03-06T10:02:12.887 DEBUG:teuthology.parallel:result is None 2026-03-06T10:02:12.888 INFO:teuthology.orchestra.run.vm01.stdout:107374182404 2026-03-06T10:02:12.903 INFO:teuthology.orchestra.run.vm01.stdout:68719476743 2026-03-06T10:02:12.954 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:12 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/737946751' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-06T10:02:12.954 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:12 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/1068887717' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-06T10:02:12.954 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:12 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/3739639913' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-06T10:02:12.954 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:12 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/3918555304' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-06T10:02:12.954 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:12 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/2781474966' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-06T10:02:13.054 INFO:tasks.cephadm.ceph_manager.ceph:need seq 90194313221 got 90194313221 for osd.2 2026-03-06T10:02:13.054 DEBUG:teuthology.parallel:result is None 2026-03-06T10:02:13.120 INFO:teuthology.orchestra.run.vm01.stdout:107374182405 2026-03-06T10:02:13.146 INFO:tasks.cephadm.ceph_manager.ceph:need seq 68719476742 got 68719476743 for osd.0 2026-03-06T10:02:13.146 DEBUG:teuthology.parallel:result is None 2026-03-06T10:02:13.159 INFO:tasks.cephadm.ceph_manager.ceph:need seq 107374182404 got 107374182404 for osd.4 2026-03-06T10:02:13.159 DEBUG:teuthology.parallel:result is None 2026-03-06T10:02:13.184 INFO:teuthology.orchestra.run.vm01.stdout:115964116996 2026-03-06T10:02:13.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:12 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/737946751' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-06T10:02:13.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:12 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/1068887717' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-06T10:02:13.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:12 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/3739639913' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-06T10:02:13.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:12 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/3918555304' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-06T10:02:13.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:12 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/2781474966' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-06T10:02:13.311 INFO:tasks.cephadm.ceph_manager.ceph:need seq 107374182404 got 107374182405 for osd.5 2026-03-06T10:02:13.311 DEBUG:teuthology.parallel:result is None 2026-03-06T10:02:13.352 INFO:tasks.cephadm.ceph_manager.ceph:need seq 115964116996 got 115964116996 for osd.7 2026-03-06T10:02:13.352 DEBUG:teuthology.parallel:result is None 2026-03-06T10:02:13.352 INFO:tasks.cephadm.ceph_manager.ceph:waiting for clean 2026-03-06T10:02:13.352 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph pg dump --format=json 2026-03-06T10:02:13.673 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:13.986 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:02:13.986 INFO:teuthology.orchestra.run.vm01.stderr:dumped all 2026-03-06T10:02:14.133 INFO:teuthology.orchestra.run.vm01.stdout:{"pg_ready":true,"pg_map":{"version":59,"stamp":"2026-03-06T09:02:12.825835+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":3,"num_osds":8,"num_per_pool_osds":8,"num_per_pool_omap_osds":3,"kb":167739392,"kb_used":218272,"kb_used_data":3532,"kb_used_omap":12,"kb_used_meta":214515,"kb_avail":167521120,"statfs":{"total":171765137408,"available":171541626880,"internally_reserved":0,"allocated":3616768,"data_stored":2220872,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":12712,"internal_metadata":219663960},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"11.489532"},"pg_stats":[{"pgid":"1.0","version":"23'32","reported_seq":66,"reported_epoch":29,"state":"active+clean","last_fresh":"2026-03-06T09:02:01.442077+0000","last_change":"2026-03-06T09:01:54.245572+0000","last_active":"2026-03-06T09:02:01.442077+0000","last_peered":"2026-03-06T09:02:01.442077+0000","last_clean":"2026-03-06T09:02:01.442077+0000","last_became_active":"2026-03-06T09:01:54.245430+0000","last_became_peered":"2026-03-06T09:01:54.245430+0000","last_unstale":"2026-03-06T09:02:01.442077+0000","last_undegraded":"2026-03-06T09:02:01.442077+0000","last_fullsized":"2026-03-06T09:02:01.442077+0000","mapping_epoch":22,"log_start":"0'0","ondisk_log_start":"0'0","created":22,"last_epoch_clean":23,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-06T09:01:53.063109+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-06T09:01:53.063109+0000","last_clean_scrub_stamp":"2026-03-06T09:01:53.063109+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-07T14:19:02.605605+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,0,2],"acting":[1,0,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1388544,"data_stored":1377840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":3}],"osd_stats":[{"osd":6,"up_from":28,"seq":120259084292,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27112,"kb_used_data":272,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940312,"statfs":{"total":21470642176,"available":21442879488,"internally_reserved":0,"allocated":278528,"data_stored":105379,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1589,"internal_metadata":27457995},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":7,"up_from":27,"seq":115964116996,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27116,"kb_used_data":272,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940308,"statfs":{"total":21470642176,"available":21442875392,"internally_reserved":0,"allocated":278528,"data_stored":105379,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1585,"internal_metadata":27457999},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":5,"up_from":25,"seq":107374182405,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27116,"kb_used_data":272,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940308,"statfs":{"total":21470642176,"available":21442875392,"internally_reserved":0,"allocated":278528,"data_stored":105379,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":25,"seq":107374182405,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27112,"kb_used_data":272,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940312,"statfs":{"total":21470642176,"available":21442879488,"internally_reserved":0,"allocated":278528,"data_stored":105379,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":21,"seq":90194313222,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27564,"kb_used_data":724,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939860,"statfs":{"total":21470642176,"available":21442416640,"internally_reserved":0,"allocated":741376,"data_stored":564659,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":20,"seq":85899345926,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27120,"kb_used_data":272,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940304,"statfs":{"total":21470642176,"available":21442871296,"internally_reserved":0,"allocated":278528,"data_stored":105379,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1588,"internal_metadata":27457996},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":17,"seq":73014444038,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27564,"kb_used_data":724,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939860,"statfs":{"total":21470642176,"available":21442416640,"internally_reserved":0,"allocated":741376,"data_stored":564659,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":16,"seq":68719476743,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27568,"kb_used_data":724,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939856,"statfs":{"total":21470642176,"available":21442412544,"internally_reserved":0,"allocated":741376,"data_stored":564659,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-06T10:02:14.133 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph pg dump --format=json 2026-03-06T10:02:14.153 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:13 vm01 ceph-mon[50931]: pgmap v59: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-06T10:02:14.153 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:13 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/3431849536' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-06T10:02:14.153 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:13 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/306764278' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-06T10:02:14.153 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:13 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/3153086896' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-06T10:02:14.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:13 vm07 ceph-mon[55141]: pgmap v59: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-06T10:02:14.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:13 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/3431849536' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-06T10:02:14.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:13 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/306764278' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-06T10:02:14.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:13 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/3153086896' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-06T10:02:14.451 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:14.760 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:02:14.760 INFO:teuthology.orchestra.run.vm01.stderr:dumped all 2026-03-06T10:02:14.937 INFO:teuthology.orchestra.run.vm01.stdout:{"pg_ready":true,"pg_map":{"version":59,"stamp":"2026-03-06T09:02:12.825835+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":3,"num_osds":8,"num_per_pool_osds":8,"num_per_pool_omap_osds":3,"kb":167739392,"kb_used":218272,"kb_used_data":3532,"kb_used_omap":12,"kb_used_meta":214515,"kb_avail":167521120,"statfs":{"total":171765137408,"available":171541626880,"internally_reserved":0,"allocated":3616768,"data_stored":2220872,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":12712,"internal_metadata":219663960},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"11.489532"},"pg_stats":[{"pgid":"1.0","version":"23'32","reported_seq":66,"reported_epoch":29,"state":"active+clean","last_fresh":"2026-03-06T09:02:01.442077+0000","last_change":"2026-03-06T09:01:54.245572+0000","last_active":"2026-03-06T09:02:01.442077+0000","last_peered":"2026-03-06T09:02:01.442077+0000","last_clean":"2026-03-06T09:02:01.442077+0000","last_became_active":"2026-03-06T09:01:54.245430+0000","last_became_peered":"2026-03-06T09:01:54.245430+0000","last_unstale":"2026-03-06T09:02:01.442077+0000","last_undegraded":"2026-03-06T09:02:01.442077+0000","last_fullsized":"2026-03-06T09:02:01.442077+0000","mapping_epoch":22,"log_start":"0'0","ondisk_log_start":"0'0","created":22,"last_epoch_clean":23,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-06T09:01:53.063109+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-06T09:01:53.063109+0000","last_clean_scrub_stamp":"2026-03-06T09:01:53.063109+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-07T14:19:02.605605+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,0,2],"acting":[1,0,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1388544,"data_stored":1377840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":3}],"osd_stats":[{"osd":6,"up_from":28,"seq":120259084292,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27112,"kb_used_data":272,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940312,"statfs":{"total":21470642176,"available":21442879488,"internally_reserved":0,"allocated":278528,"data_stored":105379,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1589,"internal_metadata":27457995},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":7,"up_from":27,"seq":115964116996,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27116,"kb_used_data":272,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940308,"statfs":{"total":21470642176,"available":21442875392,"internally_reserved":0,"allocated":278528,"data_stored":105379,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1585,"internal_metadata":27457999},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":5,"up_from":25,"seq":107374182405,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27116,"kb_used_data":272,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940308,"statfs":{"total":21470642176,"available":21442875392,"internally_reserved":0,"allocated":278528,"data_stored":105379,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":25,"seq":107374182405,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27112,"kb_used_data":272,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940312,"statfs":{"total":21470642176,"available":21442879488,"internally_reserved":0,"allocated":278528,"data_stored":105379,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":21,"seq":90194313222,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27564,"kb_used_data":724,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939860,"statfs":{"total":21470642176,"available":21442416640,"internally_reserved":0,"allocated":741376,"data_stored":564659,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":20,"seq":85899345926,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27120,"kb_used_data":272,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940304,"statfs":{"total":21470642176,"available":21442871296,"internally_reserved":0,"allocated":278528,"data_stored":105379,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1588,"internal_metadata":27457996},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":17,"seq":73014444038,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27564,"kb_used_data":724,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939860,"statfs":{"total":21470642176,"available":21442416640,"internally_reserved":0,"allocated":741376,"data_stored":564659,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":16,"seq":68719476743,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27568,"kb_used_data":724,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939856,"statfs":{"total":21470642176,"available":21442412544,"internally_reserved":0,"allocated":741376,"data_stored":564659,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-06T10:02:14.938 INFO:tasks.cephadm.ceph_manager.ceph:clean! 2026-03-06T10:02:14.938 INFO:tasks.ceph:Waiting until ceph cluster ceph is healthy... 2026-03-06T10:02:14.938 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy 2026-03-06T10:02:14.938 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- ceph health --format=json 2026-03-06T10:02:15.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:14 vm01 ceph-mon[50931]: from='client.14536 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T10:02:15.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:14 vm07 ceph-mon[55141]: from='client.14536 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T10:02:15.294 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:15.642 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T10:02:15.642 INFO:teuthology.orchestra.run.vm01.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-03-06T10:02:15.819 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy done 2026-03-06T10:02:15.820 INFO:tasks.cephadm:Setup complete, yielding 2026-03-06T10:02:15.820 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-06T10:02:15.822 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm01.local 2026-03-06T10:02:15.822 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- bash -c 'ceph orch status' 2026-03-06T10:02:15.912 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:15 vm01 ceph-mon[50931]: from='client.14540 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T10:02:15.912 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:15 vm01 ceph-mon[50931]: pgmap v60: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-06T10:02:16.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:15 vm07 ceph-mon[55141]: from='client.14540 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T10:02:16.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:15 vm07 ceph-mon[55141]: pgmap v60: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-06T10:02:16.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:15 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/2318331269' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-06T10:02:16.201 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:16.227 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:15 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/2318331269' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-06T10:02:16.551 INFO:teuthology.orchestra.run.vm01.stdout:Backend: cephadm 2026-03-06T10:02:16.551 INFO:teuthology.orchestra.run.vm01.stdout:Available: Yes 2026-03-06T10:02:16.551 INFO:teuthology.orchestra.run.vm01.stdout:Paused: No 2026-03-06T10:02:16.725 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- bash -c 'ceph orch ps' 2026-03-06T10:02:17.055 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:17.386 INFO:teuthology.orchestra.run.vm01.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-06T10:02:17.386 INFO:teuthology.orchestra.run.vm01.stdout:alertmanager.vm01 vm01 *:9093,9094 running (63s) 17s ago 113s 24.7M - 0.25.0 c8568f914cd2 ae3e783beb20 2026-03-06T10:02:17.386 INFO:teuthology.orchestra.run.vm01.stdout:ceph-exporter.vm01 vm01 running (2m) 17s ago 2m 8732k - 19.2.3-47-gc24117fd552 306e97de47e9 532db93cc68d 2026-03-06T10:02:17.386 INFO:teuthology.orchestra.run.vm01.stdout:ceph-exporter.vm07 vm07 running (81s) 18s ago 81s 6853k - 19.2.3-47-gc24117fd552 306e97de47e9 cbce26c50852 2026-03-06T10:02:17.386 INFO:teuthology.orchestra.run.vm01.stdout:crash.vm01 vm01 running (2m) 17s ago 2m 11.1M - 19.2.3-47-gc24117fd552 306e97de47e9 008b4b1e5344 2026-03-06T10:02:17.386 INFO:teuthology.orchestra.run.vm01.stdout:crash.vm07 vm07 running (80s) 18s ago 80s 11.2M - 19.2.3-47-gc24117fd552 306e97de47e9 13f1fe511623 2026-03-06T10:02:17.386 INFO:teuthology.orchestra.run.vm01.stdout:grafana.vm01 vm01 *:3000 running (61s) 17s ago 106s 69.6M - 10.4.0 c8b91775d855 20552f8840fd 2026-03-06T10:02:17.386 INFO:teuthology.orchestra.run.vm01.stdout:mgr.vm01.ifwqbh vm01 *:9283,8765,8443 running (3m) 17s ago 3m 546M - 19.2.3-47-gc24117fd552 306e97de47e9 a4e425c7ce03 2026-03-06T10:02:17.386 INFO:teuthology.orchestra.run.vm01.stdout:mgr.vm07.myglqt vm07 *:8443,9283,8765 running (75s) 18s ago 75s 476M - 19.2.3-47-gc24117fd552 306e97de47e9 c0f2e91f8484 2026-03-06T10:02:17.386 INFO:teuthology.orchestra.run.vm01.stdout:mon.vm01 vm01 running (3m) 17s ago 3m 51.0M 2048M 19.2.3-47-gc24117fd552 306e97de47e9 f5dcc017bf81 2026-03-06T10:02:17.386 INFO:teuthology.orchestra.run.vm01.stdout:mon.vm07 vm07 running (74s) 18s ago 73s 42.4M 2048M 19.2.3-47-gc24117fd552 306e97de47e9 e7653b40e484 2026-03-06T10:02:17.386 INFO:teuthology.orchestra.run.vm01.stdout:node-exporter.vm01 vm01 *:9100 running (117s) 17s ago 117s 9517k - 1.7.0 72c9c2088986 89973855e80f 2026-03-06T10:02:17.386 INFO:teuthology.orchestra.run.vm01.stdout:node-exporter.vm07 vm07 *:9100 running (76s) 18s ago 76s 9156k - 1.7.0 72c9c2088986 7dd625d6098a 2026-03-06T10:02:17.386 INFO:teuthology.orchestra.run.vm01.stdout:osd.0 vm07 running (34s) 18s ago 34s 54.9M 4096M 19.2.3-47-gc24117fd552 306e97de47e9 d262919f24c0 2026-03-06T10:02:17.386 INFO:teuthology.orchestra.run.vm01.stdout:osd.1 vm01 running (33s) 17s ago 33s 39.7M 4096M 19.2.3-47-gc24117fd552 306e97de47e9 9f94e77f86c2 2026-03-06T10:02:17.386 INFO:teuthology.orchestra.run.vm01.stdout:osd.2 vm01 running (29s) 17s ago 29s 33.5M 4096M 19.2.3-47-gc24117fd552 306e97de47e9 1c4bb387c2f3 2026-03-06T10:02:17.386 INFO:teuthology.orchestra.run.vm01.stdout:osd.3 vm07 running (30s) 18s ago 30s 31.0M 4096M 19.2.3-47-gc24117fd552 306e97de47e9 dee5ea9abded 2026-03-06T10:02:17.386 INFO:teuthology.orchestra.run.vm01.stdout:osd.4 vm01 running (25s) 17s ago 25s 56.4M 4096M 19.2.3-47-gc24117fd552 306e97de47e9 641c377a0cce 2026-03-06T10:02:17.386 INFO:teuthology.orchestra.run.vm01.stdout:osd.5 vm07 running (26s) 18s ago 26s 52.4M 4096M 19.2.3-47-gc24117fd552 306e97de47e9 26b3cbc44f66 2026-03-06T10:02:17.386 INFO:teuthology.orchestra.run.vm01.stdout:osd.6 vm01 running (21s) 17s ago 21s 25.3M 4096M 19.2.3-47-gc24117fd552 306e97de47e9 2f1518688aa6 2026-03-06T10:02:17.387 INFO:teuthology.orchestra.run.vm01.stdout:osd.7 vm07 running (23s) 18s ago 23s 52.0M 4096M 19.2.3-47-gc24117fd552 306e97de47e9 80f630c48863 2026-03-06T10:02:17.387 INFO:teuthology.orchestra.run.vm01.stdout:prometheus.vm01 vm01 *:9095 running (60s) 17s ago 100s 36.1M - 2.51.0 1d3b7f56885b 71b403c178cf 2026-03-06T10:02:17.544 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- bash -c 'ceph orch ls' 2026-03-06T10:02:17.872 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:18.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:17 vm07 ceph-mon[55141]: from='client.14548 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:18.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:17 vm07 ceph-mon[55141]: pgmap v61: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-06T10:02:18.203 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:17 vm01 ceph-mon[50931]: from='client.14548 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:18.203 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:17 vm01 ceph-mon[50931]: pgmap v61: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-06T10:02:18.204 INFO:teuthology.orchestra.run.vm01.stdout:NAME PORTS RUNNING REFRESHED AGE PLACEMENT 2026-03-06T10:02:18.204 INFO:teuthology.orchestra.run.vm01.stdout:alertmanager ?:9093,9094 1/1 18s ago 2m count:1 2026-03-06T10:02:18.204 INFO:teuthology.orchestra.run.vm01.stdout:ceph-exporter 2/2 19s ago 2m * 2026-03-06T10:02:18.204 INFO:teuthology.orchestra.run.vm01.stdout:crash 2/2 19s ago 2m * 2026-03-06T10:02:18.204 INFO:teuthology.orchestra.run.vm01.stdout:grafana ?:3000 1/1 18s ago 2m count:1 2026-03-06T10:02:18.204 INFO:teuthology.orchestra.run.vm01.stdout:mgr 2/2 19s ago 2m count:2 2026-03-06T10:02:18.204 INFO:teuthology.orchestra.run.vm01.stdout:mon 2/2 19s ago 117s vm01:192.168.123.101=vm01;vm07:192.168.123.107=vm07;count:2 2026-03-06T10:02:18.204 INFO:teuthology.orchestra.run.vm01.stdout:node-exporter ?:9100 2/2 19s ago 2m * 2026-03-06T10:02:18.204 INFO:teuthology.orchestra.run.vm01.stdout:osd.all-available-devices 8 19s ago 65s * 2026-03-06T10:02:18.204 INFO:teuthology.orchestra.run.vm01.stdout:prometheus ?:9095 1/1 18s ago 2m count:1 2026-03-06T10:02:18.360 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- bash -c 'ceph orch host ls' 2026-03-06T10:02:18.682 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:19.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:18 vm01 ceph-mon[50931]: from='client.14552 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:19.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:18 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T10:02:19.034 INFO:teuthology.orchestra.run.vm01.stdout:HOST ADDR LABELS STATUS 2026-03-06T10:02:19.034 INFO:teuthology.orchestra.run.vm01.stdout:vm01 192.168.123.101 2026-03-06T10:02:19.034 INFO:teuthology.orchestra.run.vm01.stdout:vm07 192.168.123.107 2026-03-06T10:02:19.034 INFO:teuthology.orchestra.run.vm01.stdout:2 hosts in cluster 2026-03-06T10:02:19.188 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- bash -c 'ceph orch device ls' 2026-03-06T10:02:19.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:18 vm07 ceph-mon[55141]: from='client.14552 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:19.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:18 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T10:02:19.517 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:19.851 INFO:teuthology.orchestra.run.vm01.stdout:HOST PATH TYPE DEVICE ID SIZE AVAILABLE REFRESHED REJECT REASONS 2026-03-06T10:02:19.851 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 17s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-06T10:02:19.851 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/vdb hdd DWNBRSTVMM01001 20.0G No 17s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T10:02:19.851 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/vdc hdd DWNBRSTVMM01002 20.0G No 17s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T10:02:19.851 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/vdd hdd DWNBRSTVMM01003 20.0G No 17s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T10:02:19.851 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/vde hdd DWNBRSTVMM01004 20.0G No 17s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T10:02:19.851 INFO:teuthology.orchestra.run.vm01.stdout:vm07 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 18s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-06T10:02:19.851 INFO:teuthology.orchestra.run.vm01.stdout:vm07 /dev/vdb hdd DWNBRSTVMM07001 20.0G No 18s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T10:02:19.851 INFO:teuthology.orchestra.run.vm01.stdout:vm07 /dev/vdc hdd DWNBRSTVMM07002 20.0G No 18s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T10:02:19.851 INFO:teuthology.orchestra.run.vm01.stdout:vm07 /dev/vdd hdd DWNBRSTVMM07003 20.0G No 18s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T10:02:19.851 INFO:teuthology.orchestra.run.vm01.stdout:vm07 /dev/vde hdd DWNBRSTVMM07004 20.0G No 18s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T10:02:20.036 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- bash -c 'ceph orch ls | grep '"'"'^osd.all-available-devices '"'"'' 2026-03-06T10:02:20.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:19 vm07 ceph-mon[55141]: from='client.14556 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:20.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:19 vm07 ceph-mon[55141]: pgmap v62: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-06T10:02:20.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:19 vm07 ceph-mon[55141]: from='client.14560 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:20.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:19 vm01 ceph-mon[50931]: from='client.14556 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:20.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:19 vm01 ceph-mon[50931]: pgmap v62: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-06T10:02:20.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:19 vm01 ceph-mon[50931]: from='client.14560 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:20.364 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:20.708 INFO:teuthology.orchestra.run.vm01.stdout:osd.all-available-devices 8 21s ago 68s * 2026-03-06T10:02:20.867 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-06T10:02:20.869 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm01.local 2026-03-06T10:02:20.869 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3ab6672-193a-11f1-b81f-a119763c7190 -- bash -c 'set -e 2026-03-06T10:02:20.869 DEBUG:teuthology.orchestra.run.vm01:> set -x 2026-03-06T10:02:20.869 DEBUG:teuthology.orchestra.run.vm01:> ceph orch ps 2026-03-06T10:02:20.870 DEBUG:teuthology.orchestra.run.vm01:> ceph orch device ls 2026-03-06T10:02:20.870 DEBUG:teuthology.orchestra.run.vm01:> DEVID=$(ceph device ls | grep osd.1 | awk '"'"'{print $1}'"'"') 2026-03-06T10:02:20.870 DEBUG:teuthology.orchestra.run.vm01:> HOST=$(ceph orch device ls | grep "$DEVID" | awk '"'"'{print $1}'"'"') 2026-03-06T10:02:20.870 DEBUG:teuthology.orchestra.run.vm01:> DEV=$(ceph orch device ls | grep "$DEVID" | awk '"'"'{print $2}'"'"') 2026-03-06T10:02:20.870 DEBUG:teuthology.orchestra.run.vm01:> echo "host $HOST, dev $DEV, devid $DEVID" 2026-03-06T10:02:20.870 DEBUG:teuthology.orchestra.run.vm01:> ceph orch osd rm --zap --replace 1 2026-03-06T10:02:20.870 DEBUG:teuthology.orchestra.run.vm01:> while ceph orch osd rm status | grep ^1 ; do sleep 5 ; done 2026-03-06T10:02:20.870 DEBUG:teuthology.orchestra.run.vm01:> while ! ceph osd dump | grep osd.1 | grep "up\s*in" ; do sleep 5 ; done 2026-03-06T10:02:20.870 DEBUG:teuthology.orchestra.run.vm01:> ' 2026-03-06T10:02:21.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:20 vm01 ceph-mon[50931]: from='client.14564 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:21.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:20 vm07 ceph-mon[55141]: from='client.14564 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:21.208 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/mon.vm01/config 2026-03-06T10:02:21.288 INFO:teuthology.orchestra.run.vm01.stderr:+ ceph orch ps 2026-03-06T10:02:21.538 INFO:teuthology.orchestra.run.vm01.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-06T10:02:21.538 INFO:teuthology.orchestra.run.vm01.stdout:alertmanager.vm01 vm01 *:9093,9094 running (67s) 21s ago 117s 24.7M - 0.25.0 c8568f914cd2 ae3e783beb20 2026-03-06T10:02:21.538 INFO:teuthology.orchestra.run.vm01.stdout:ceph-exporter.vm01 vm01 running (2m) 21s ago 2m 8732k - 19.2.3-47-gc24117fd552 306e97de47e9 532db93cc68d 2026-03-06T10:02:21.539 INFO:teuthology.orchestra.run.vm01.stdout:ceph-exporter.vm07 vm07 running (85s) 22s ago 85s 6853k - 19.2.3-47-gc24117fd552 306e97de47e9 cbce26c50852 2026-03-06T10:02:21.539 INFO:teuthology.orchestra.run.vm01.stdout:crash.vm01 vm01 running (2m) 21s ago 2m 11.1M - 19.2.3-47-gc24117fd552 306e97de47e9 008b4b1e5344 2026-03-06T10:02:21.539 INFO:teuthology.orchestra.run.vm01.stdout:crash.vm07 vm07 running (84s) 22s ago 84s 11.2M - 19.2.3-47-gc24117fd552 306e97de47e9 13f1fe511623 2026-03-06T10:02:21.539 INFO:teuthology.orchestra.run.vm01.stdout:grafana.vm01 vm01 *:3000 running (65s) 21s ago 110s 69.6M - 10.4.0 c8b91775d855 20552f8840fd 2026-03-06T10:02:21.539 INFO:teuthology.orchestra.run.vm01.stdout:mgr.vm01.ifwqbh vm01 *:9283,8765,8443 running (3m) 21s ago 3m 546M - 19.2.3-47-gc24117fd552 306e97de47e9 a4e425c7ce03 2026-03-06T10:02:21.539 INFO:teuthology.orchestra.run.vm01.stdout:mgr.vm07.myglqt vm07 *:8443,9283,8765 running (79s) 22s ago 79s 476M - 19.2.3-47-gc24117fd552 306e97de47e9 c0f2e91f8484 2026-03-06T10:02:21.539 INFO:teuthology.orchestra.run.vm01.stdout:mon.vm01 vm01 running (3m) 21s ago 3m 51.0M 2048M 19.2.3-47-gc24117fd552 306e97de47e9 f5dcc017bf81 2026-03-06T10:02:21.539 INFO:teuthology.orchestra.run.vm01.stdout:mon.vm07 vm07 running (78s) 22s ago 78s 42.4M 2048M 19.2.3-47-gc24117fd552 306e97de47e9 e7653b40e484 2026-03-06T10:02:21.539 INFO:teuthology.orchestra.run.vm01.stdout:node-exporter.vm01 vm01 *:9100 running (2m) 21s ago 2m 9517k - 1.7.0 72c9c2088986 89973855e80f 2026-03-06T10:02:21.539 INFO:teuthology.orchestra.run.vm01.stdout:node-exporter.vm07 vm07 *:9100 running (81s) 22s ago 81s 9156k - 1.7.0 72c9c2088986 7dd625d6098a 2026-03-06T10:02:21.539 INFO:teuthology.orchestra.run.vm01.stdout:osd.0 vm07 running (38s) 22s ago 38s 54.9M 4096M 19.2.3-47-gc24117fd552 306e97de47e9 d262919f24c0 2026-03-06T10:02:21.539 INFO:teuthology.orchestra.run.vm01.stdout:osd.1 vm01 running (38s) 21s ago 37s 39.7M 4096M 19.2.3-47-gc24117fd552 306e97de47e9 9f94e77f86c2 2026-03-06T10:02:21.539 INFO:teuthology.orchestra.run.vm01.stdout:osd.2 vm01 running (33s) 21s ago 33s 33.5M 4096M 19.2.3-47-gc24117fd552 306e97de47e9 1c4bb387c2f3 2026-03-06T10:02:21.539 INFO:teuthology.orchestra.run.vm01.stdout:osd.3 vm07 running (34s) 22s ago 34s 31.0M 4096M 19.2.3-47-gc24117fd552 306e97de47e9 dee5ea9abded 2026-03-06T10:02:21.539 INFO:teuthology.orchestra.run.vm01.stdout:osd.4 vm01 running (29s) 21s ago 29s 56.4M 4096M 19.2.3-47-gc24117fd552 306e97de47e9 641c377a0cce 2026-03-06T10:02:21.539 INFO:teuthology.orchestra.run.vm01.stdout:osd.5 vm07 running (31s) 22s ago 31s 52.4M 4096M 19.2.3-47-gc24117fd552 306e97de47e9 26b3cbc44f66 2026-03-06T10:02:21.539 INFO:teuthology.orchestra.run.vm01.stdout:osd.6 vm01 running (25s) 21s ago 25s 25.3M 4096M 19.2.3-47-gc24117fd552 306e97de47e9 2f1518688aa6 2026-03-06T10:02:21.539 INFO:teuthology.orchestra.run.vm01.stdout:osd.7 vm07 running (27s) 22s ago 27s 52.0M 4096M 19.2.3-47-gc24117fd552 306e97de47e9 80f630c48863 2026-03-06T10:02:21.539 INFO:teuthology.orchestra.run.vm01.stdout:prometheus.vm01 vm01 *:9095 running (64s) 21s ago 104s 36.1M - 2.51.0 1d3b7f56885b 71b403c178cf 2026-03-06T10:02:21.549 INFO:teuthology.orchestra.run.vm01.stderr:+ ceph orch device ls 2026-03-06T10:02:21.692 INFO:teuthology.orchestra.run.vm01.stdout:HOST PATH TYPE DEVICE ID SIZE AVAILABLE REFRESHED REJECT REASONS 2026-03-06T10:02:21.692 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 19s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-06T10:02:21.692 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/vdb hdd DWNBRSTVMM01001 20.0G No 19s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T10:02:21.692 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/vdc hdd DWNBRSTVMM01002 20.0G No 19s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T10:02:21.692 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/vdd hdd DWNBRSTVMM01003 20.0G No 19s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T10:02:21.692 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/vde hdd DWNBRSTVMM01004 20.0G No 19s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T10:02:21.692 INFO:teuthology.orchestra.run.vm01.stdout:vm07 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 20s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-06T10:02:21.692 INFO:teuthology.orchestra.run.vm01.stdout:vm07 /dev/vdb hdd DWNBRSTVMM07001 20.0G No 20s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T10:02:21.692 INFO:teuthology.orchestra.run.vm01.stdout:vm07 /dev/vdc hdd DWNBRSTVMM07002 20.0G No 20s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T10:02:21.692 INFO:teuthology.orchestra.run.vm01.stdout:vm07 /dev/vdd hdd DWNBRSTVMM07003 20.0G No 20s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T10:02:21.692 INFO:teuthology.orchestra.run.vm01.stdout:vm07 /dev/vde hdd DWNBRSTVMM07004 20.0G No 20s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T10:02:21.700 INFO:teuthology.orchestra.run.vm01.stderr:++ ceph device ls 2026-03-06T10:02:21.700 INFO:teuthology.orchestra.run.vm01.stderr:++ awk '{print $1}' 2026-03-06T10:02:21.702 INFO:teuthology.orchestra.run.vm01.stderr:++ grep osd.1 2026-03-06T10:02:21.848 INFO:teuthology.orchestra.run.vm01.stderr:+ DEVID=DWNBRSTVMM01001 2026-03-06T10:02:21.849 INFO:teuthology.orchestra.run.vm01.stderr:++ awk '{print $1}' 2026-03-06T10:02:21.849 INFO:teuthology.orchestra.run.vm01.stderr:++ ceph orch device ls 2026-03-06T10:02:21.849 INFO:teuthology.orchestra.run.vm01.stderr:++ grep DWNBRSTVMM01001 2026-03-06T10:02:22.009 INFO:teuthology.orchestra.run.vm01.stderr:+ HOST=vm01 2026-03-06T10:02:22.009 INFO:teuthology.orchestra.run.vm01.stderr:++ ceph orch device ls 2026-03-06T10:02:22.009 INFO:teuthology.orchestra.run.vm01.stderr:++ awk '{print $2}' 2026-03-06T10:02:22.011 INFO:teuthology.orchestra.run.vm01.stderr:++ grep DWNBRSTVMM01001 2026-03-06T10:02:22.168 INFO:teuthology.orchestra.run.vm01.stderr:+ DEV=/dev/vdb 2026-03-06T10:02:22.168 INFO:teuthology.orchestra.run.vm01.stderr:+ echo 'host vm01, dev /dev/vdb, devid DWNBRSTVMM01001' 2026-03-06T10:02:22.168 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:21 vm01 ceph-mon[50931]: from='client.14568 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:22.168 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:21 vm01 ceph-mon[50931]: pgmap v63: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-06T10:02:22.169 INFO:teuthology.orchestra.run.vm01.stdout:host vm01, dev /dev/vdb, devid DWNBRSTVMM01001 2026-03-06T10:02:22.169 INFO:teuthology.orchestra.run.vm01.stderr:+ ceph orch osd rm --zap --replace 1 2026-03-06T10:02:22.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:21 vm07 ceph-mon[55141]: from='client.14568 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:22.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:21 vm07 ceph-mon[55141]: pgmap v63: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-06T10:02:22.319 INFO:teuthology.orchestra.run.vm01.stdout:Scheduled OSD(s) for removal. 2026-03-06T10:02:22.336 INFO:teuthology.orchestra.run.vm01.stderr:+ grep '^1' 2026-03-06T10:02:22.337 INFO:teuthology.orchestra.run.vm01.stderr:+ ceph orch osd rm status 2026-03-06T10:02:22.510 INFO:teuthology.orchestra.run.vm01.stdout:1 vm01 started 1 True False True 2026-03-06T10:02:22.510 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-06T10:02:23.023 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:22 vm07 ceph-mon[55141]: from='client.14572 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:23.023 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:22 vm07 ceph-mon[55141]: from='client.14576 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:23.023 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:22 vm07 ceph-mon[55141]: from='client.14580 -' entity='client.admin' cmd=[{"prefix": "device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:23.023 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:22 vm07 ceph-mon[55141]: from='client.14584 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:23.023 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:22 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd crush tree", "format": "json"}]: dispatch 2026-03-06T10:02:23.023 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:22 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T10:02:23.023 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:22 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:23.183 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:22 vm01 ceph-mon[50931]: from='client.14572 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:23.184 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:22 vm01 ceph-mon[50931]: from='client.14576 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:23.184 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:22 vm01 ceph-mon[50931]: from='client.14580 -' entity='client.admin' cmd=[{"prefix": "device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:23.184 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:22 vm01 ceph-mon[50931]: from='client.14584 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:23.184 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:22 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd crush tree", "format": "json"}]: dispatch 2026-03-06T10:02:23.184 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:22 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T10:02:23.184 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:22 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:24.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:23 vm01 ceph-a3ab6672-193a-11f1-b81f-a119763c7190-mon-vm01[50907]: 2026-03-06T09:02:23.720+0000 7fba00b27640 -1 mon.vm01@0(leader).osd e29 definitely_dead 0 2026-03-06T10:02:24.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:23 vm01 ceph-mon[50931]: from='client.14588 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:24.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:23 vm01 ceph-mon[50931]: from='client.14592 -' entity='client.admin' cmd=[{"prefix": "orch osd rm", "zap": true, "replace": true, "osd_id": ["1"], "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:24.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:23 vm01 ceph-mon[50931]: osd.1 crush weight is 0.0194854736328125 2026-03-06T10:02:24.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:23 vm01 ceph-mon[50931]: from='client.14596 -' entity='client.admin' cmd=[{"prefix": "orch osd rm status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:24.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:23 vm01 ceph-mon[50931]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:24.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:23 vm01 ceph-mon[50931]: pgmap v64: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-06T10:02:24.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:23 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:24.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:23 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:24.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:23 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:24.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:23 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:24.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:23 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:02:24.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:23 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T10:02:24.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:23 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:24.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:23 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd ok-to-stop", "ids": ["1"]}]: dispatch 2026-03-06T10:02:24.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:23 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd out", "ids": ["1"]}]: dispatch 2026-03-06T10:02:24.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:23 vm01 ceph-mon[50931]: Client mgr.vm01.ifwqbh marked osd.1 out, while it was still marked up 2026-03-06T10:02:24.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:23 vm07 ceph-mon[55141]: from='client.14588 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:24.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:23 vm07 ceph-mon[55141]: from='client.14592 -' entity='client.admin' cmd=[{"prefix": "orch osd rm", "zap": true, "replace": true, "osd_id": ["1"], "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:24.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:23 vm07 ceph-mon[55141]: osd.1 crush weight is 0.0194854736328125 2026-03-06T10:02:24.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:23 vm07 ceph-mon[55141]: from='client.14596 -' entity='client.admin' cmd=[{"prefix": "orch osd rm status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:24.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:23 vm07 ceph-mon[55141]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:24.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:23 vm07 ceph-mon[55141]: pgmap v64: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-06T10:02:24.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:23 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:24.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:23 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:24.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:23 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:24.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:23 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:24.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:23 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:02:24.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:23 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T10:02:24.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:23 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:24.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:23 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd ok-to-stop", "ids": ["1"]}]: dispatch 2026-03-06T10:02:24.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:23 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd out", "ids": ["1"]}]: dispatch 2026-03-06T10:02:24.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:23 vm07 ceph-mon[55141]: Client mgr.vm01.ifwqbh marked osd.1 out, while it was still marked up 2026-03-06T10:02:25.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:24 vm07 ceph-mon[55141]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:25.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:24 vm07 ceph-mon[55141]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd ok-to-stop", "ids": ["1"]}]: dispatch 2026-03-06T10:02:25.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:24 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd='[{"prefix": "osd out", "ids": ["1"]}]': finished 2026-03-06T10:02:25.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:24 vm07 ceph-mon[55141]: osdmap e30: 8 total, 8 up, 7 in 2026-03-06T10:02:25.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:24 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:25.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:24 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:25.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:24 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T10:02:25.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:24 vm01 ceph-mon[50931]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:25.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:24 vm01 ceph-mon[50931]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd ok-to-stop", "ids": ["1"]}]: dispatch 2026-03-06T10:02:25.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:24 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd='[{"prefix": "osd out", "ids": ["1"]}]': finished 2026-03-06T10:02:25.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:24 vm01 ceph-mon[50931]: osdmap e30: 8 total, 8 up, 7 in 2026-03-06T10:02:25.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:24 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:25.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:24 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:25.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:24 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T10:02:26.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:25 vm07 ceph-mon[55141]: osd.1 now out 2026-03-06T10:02:26.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:25 vm07 ceph-mon[55141]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:26.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:25 vm07 ceph-mon[55141]: pgmap v66: 1 pgs: 1 active+clean; 449 KiB data, 186 MiB used, 140 GiB / 140 GiB avail 2026-03-06T10:02:26.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:25 vm01 ceph-mon[50931]: osd.1 now out 2026-03-06T10:02:26.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:25 vm01 ceph-mon[50931]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:26.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:25 vm01 ceph-mon[50931]: pgmap v66: 1 pgs: 1 active+clean; 449 KiB data, 186 MiB used, 140 GiB / 140 GiB avail 2026-03-06T10:02:27.513 INFO:teuthology.orchestra.run.vm01.stderr:+ ceph orch osd rm status 2026-03-06T10:02:27.513 INFO:teuthology.orchestra.run.vm01.stderr:+ grep '^1' 2026-03-06T10:02:27.676 INFO:teuthology.orchestra.run.vm01.stdout:1 vm01 draining 1 True False True 2026-03-06 09:02:24.722486 2026-03-06T10:02:27.676 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-06T10:02:27.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:27 vm01 ceph-mon[50931]: osdmap e31: 8 total, 8 up, 7 in 2026-03-06T10:02:27.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:27 vm01 ceph-mon[50931]: pgmap v68: 1 pgs: 1 active+clean; 449 KiB data, 186 MiB used, 140 GiB / 140 GiB avail 2026-03-06T10:02:27.767 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:27 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:28.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:27 vm07 ceph-mon[55141]: osdmap e31: 8 total, 8 up, 7 in 2026-03-06T10:02:28.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:27 vm07 ceph-mon[55141]: pgmap v68: 1 pgs: 1 active+clean; 449 KiB data, 186 MiB used, 140 GiB / 140 GiB avail 2026-03-06T10:02:28.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:27 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:29.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:28 vm01 ceph-mon[50931]: from='client.14600 -' entity='client.admin' cmd=[{"prefix": "orch osd rm status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:29.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:28 vm01 ceph-mon[50931]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:29.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:28 vm07 ceph-mon[55141]: from='client.14600 -' entity='client.admin' cmd=[{"prefix": "orch osd rm status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:29.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:28 vm07 ceph-mon[55141]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:29.842 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:29 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T10:02:29.842 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:29 vm01 ceph-mon[50931]: pgmap v69: 1 pgs: 1 activating+degraded; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail; 2/6 objects degraded (33.333%) 2026-03-06T10:02:29.842 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:29 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:29.843 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:29 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:30.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:29 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T10:02:30.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:29 vm07 ceph-mon[55141]: pgmap v69: 1 pgs: 1 activating+degraded; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail; 2/6 objects degraded (33.333%) 2026-03-06T10:02:30.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:29 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:30.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:29 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:30.241 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:30 vm01 ceph-a3ab6672-193a-11f1-b81f-a119763c7190-mon-vm01[50907]: 2026-03-06T09:02:30.240+0000 7fba00b27640 -1 mon.vm01@0(leader).osd e31 definitely_dead 0 2026-03-06T10:02:31.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:30 vm01 ceph-mon[50931]: Health check failed: Degraded data redundancy: 2/6 objects degraded (33.333%), 1 pg degraded (PG_DEGRADED) 2026-03-06T10:02:31.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:30 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:31.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:30 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:31.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:30 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:02:31.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:30 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T10:02:31.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:30 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:31.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:30 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:31.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:30 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:31.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:30 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd safe-to-destroy", "ids": ["1"]}]: dispatch 2026-03-06T10:02:31.017 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:30 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd down", "ids": ["1"]}]: dispatch 2026-03-06T10:02:31.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:30 vm07 ceph-mon[55141]: Health check failed: Degraded data redundancy: 2/6 objects degraded (33.333%), 1 pg degraded (PG_DEGRADED) 2026-03-06T10:02:31.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:30 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:31.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:30 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:31.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:30 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:02:31.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:30 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T10:02:31.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:30 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:31.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:30 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:31.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:30 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:31.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:30 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd safe-to-destroy", "ids": ["1"]}]: dispatch 2026-03-06T10:02:31.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:30 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd down", "ids": ["1"]}]: dispatch 2026-03-06T10:02:31.877 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:31 vm01 ceph-mon[50931]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:31.877 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:31 vm01 ceph-mon[50931]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:31.877 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:31 vm01 ceph-mon[50931]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:31.877 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:31 vm01 ceph-mon[50931]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd safe-to-destroy", "ids": ["1"]}]: dispatch 2026-03-06T10:02:31.877 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:31 vm01 ceph-mon[50931]: pgmap v70: 1 pgs: 1 activating+degraded; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail; 2/6 objects degraded (33.333%) 2026-03-06T10:02:31.877 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:31 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd='[{"prefix": "osd down", "ids": ["1"]}]': finished 2026-03-06T10:02:31.877 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:31 vm01 ceph-mon[50931]: osdmap e32: 8 total, 7 up, 7 in 2026-03-06T10:02:32.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:31 vm07 ceph-mon[55141]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:32.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:31 vm07 ceph-mon[55141]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:32.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:31 vm07 ceph-mon[55141]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:32.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:31 vm07 ceph-mon[55141]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd safe-to-destroy", "ids": ["1"]}]: dispatch 2026-03-06T10:02:32.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:31 vm07 ceph-mon[55141]: pgmap v70: 1 pgs: 1 activating+degraded; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail; 2/6 objects degraded (33.333%) 2026-03-06T10:02:32.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:31 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd='[{"prefix": "osd down", "ids": ["1"]}]': finished 2026-03-06T10:02:32.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:31 vm07 ceph-mon[55141]: osdmap e32: 8 total, 7 up, 7 in 2026-03-06T10:02:32.684 INFO:teuthology.orchestra.run.vm01.stderr:+ grep '^1' 2026-03-06T10:02:32.684 INFO:teuthology.orchestra.run.vm01.stderr:+ ceph orch osd rm status 2026-03-06T10:02:32.865 INFO:teuthology.orchestra.run.vm01.stdout:1 vm01 done, waiting for purge 0 True False True 2026-03-06 09:02:24.722486 2026-03-06T10:02:32.865 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-06T10:02:32.990 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:32 vm01 ceph-mon[50931]: osd.1 now down 2026-03-06T10:02:32.990 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:32 vm01 ceph-mon[50931]: Removing daemon osd.1 from vm01 -- ports [] 2026-03-06T10:02:32.990 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:32 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth rm", "entity": "osd.1"}]: dispatch 2026-03-06T10:02:32.990 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:32 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd='[{"prefix": "auth rm", "entity": "osd.1"}]': finished 2026-03-06T10:02:32.990 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:32 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd destroy-actual", "id": 1, "yes_i_really_mean_it": true}]: dispatch 2026-03-06T10:02:32.990 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:32 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd='[{"prefix": "osd destroy-actual", "id": 1, "yes_i_really_mean_it": true}]': finished 2026-03-06T10:02:32.990 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:32 vm01 ceph-mon[50931]: osdmap e33: 8 total, 7 up, 7 in 2026-03-06T10:02:33.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:32 vm07 ceph-mon[55141]: osd.1 now down 2026-03-06T10:02:33.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:32 vm07 ceph-mon[55141]: Removing daemon osd.1 from vm01 -- ports [] 2026-03-06T10:02:33.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:32 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth rm", "entity": "osd.1"}]: dispatch 2026-03-06T10:02:33.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:32 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd='[{"prefix": "auth rm", "entity": "osd.1"}]': finished 2026-03-06T10:02:33.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:32 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd destroy-actual", "id": 1, "yes_i_really_mean_it": true}]: dispatch 2026-03-06T10:02:33.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:32 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd='[{"prefix": "osd destroy-actual", "id": 1, "yes_i_really_mean_it": true}]': finished 2026-03-06T10:02:33.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:32 vm07 ceph-mon[55141]: osdmap e33: 8 total, 7 up, 7 in 2026-03-06T10:02:34.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:33 vm01 ceph-mon[50931]: Removing key for osd.1 2026-03-06T10:02:34.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:33 vm01 ceph-mon[50931]: Successfully removed osd.1 on vm01 2026-03-06T10:02:34.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:33 vm01 ceph-mon[50931]: Successfully destroyed old osd.1 on vm01; ready for replacement 2026-03-06T10:02:34.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:33 vm01 ceph-mon[50931]: Zapping devices for osd.1 on vm01 2026-03-06T10:02:34.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:33 vm01 ceph-mon[50931]: pgmap v73: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail; 74 KiB/s, 0 objects/s recovering 2026-03-06T10:02:34.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:33 vm01 ceph-mon[50931]: from='client.24351 -' entity='client.admin' cmd=[{"prefix": "orch osd rm status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:34.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:33 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:34.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:33 vm01 ceph-mon[50931]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:34.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:33 vm07 ceph-mon[55141]: Removing key for osd.1 2026-03-06T10:02:34.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:33 vm07 ceph-mon[55141]: Successfully removed osd.1 on vm01 2026-03-06T10:02:34.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:33 vm07 ceph-mon[55141]: Successfully destroyed old osd.1 on vm01; ready for replacement 2026-03-06T10:02:34.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:33 vm07 ceph-mon[55141]: Zapping devices for osd.1 on vm01 2026-03-06T10:02:34.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:33 vm07 ceph-mon[55141]: pgmap v73: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail; 74 KiB/s, 0 objects/s recovering 2026-03-06T10:02:34.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:33 vm07 ceph-mon[55141]: from='client.24351 -' entity='client.admin' cmd=[{"prefix": "orch osd rm status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:34.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:33 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:34.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:33 vm07 ceph-mon[55141]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-06T10:02:34.765 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:34 vm01 ceph-mon[50931]: Health check cleared: PG_DEGRADED (was: Degraded data redundancy: 2/6 objects degraded (33.333%), 1 pg degraded) 2026-03-06T10:02:34.766 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:34 vm01 ceph-mon[50931]: Cluster is now healthy 2026-03-06T10:02:34.766 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:34 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T10:02:34.766 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:34 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:34.766 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:34 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T10:02:35.070 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:34 vm07 ceph-mon[55141]: Health check cleared: PG_DEGRADED (was: Degraded data redundancy: 2/6 objects degraded (33.333%), 1 pg degraded) 2026-03-06T10:02:35.071 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:34 vm07 ceph-mon[55141]: Cluster is now healthy 2026-03-06T10:02:35.071 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:34 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T10:02:35.071 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:34 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:35.071 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:34 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T10:02:36.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:35 vm01 ceph-mon[50931]: Successfully zapped devices for osd.1 on vm01 2026-03-06T10:02:36.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:35 vm01 ceph-mon[50931]: pgmap v74: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail; 56 KiB/s, 0 objects/s recovering 2026-03-06T10:02:36.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:35 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:36.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:35 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:36.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:35 vm07 ceph-mon[55141]: Successfully zapped devices for osd.1 on vm01 2026-03-06T10:02:36.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:35 vm07 ceph-mon[55141]: pgmap v74: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail; 56 KiB/s, 0 objects/s recovering 2026-03-06T10:02:36.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:35 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:36.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:35 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:37.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:36 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:37.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:36 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:37.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:36 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:37.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:36 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:37.869 INFO:teuthology.orchestra.run.vm01.stderr:+ ceph orch osd rm status 2026-03-06T10:02:37.869 INFO:teuthology.orchestra.run.vm01.stderr:+ grep '^1' 2026-03-06T10:02:38.038 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:37 vm01 ceph-mon[50931]: pgmap v75: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail; 56 KiB/s, 0 objects/s recovering 2026-03-06T10:02:38.038 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:37 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:38.038 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:37 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:38.038 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:37 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:02:38.038 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:37 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T10:02:38.038 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:37 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:38.038 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:37 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T10:02:38.038 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:37 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T10:02:38.038 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:37 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:02:38.048 INFO:teuthology.orchestra.run.vm01.stderr:+ grep osd.1 2026-03-06T10:02:38.049 INFO:teuthology.orchestra.run.vm01.stderr:+ grep 'up\s*in' 2026-03-06T10:02:38.050 INFO:teuthology.orchestra.run.vm01.stderr:+ ceph osd dump 2026-03-06T10:02:38.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:37 vm07 ceph-mon[55141]: pgmap v75: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail; 56 KiB/s, 0 objects/s recovering 2026-03-06T10:02:38.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:37 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:38.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:37 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:38.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:37 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:02:38.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:37 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T10:02:38.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:37 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:38.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:37 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T10:02:38.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:37 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T10:02:38.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:37 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:02:38.216 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-06T10:02:38.866 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:38 vm01 ceph-mon[50931]: Detected new or changed devices on vm01 2026-03-06T10:02:38.866 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:38 vm01 ceph-mon[50931]: Found osd claims -> {'vm01': ['1']} 2026-03-06T10:02:38.866 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:38 vm01 ceph-mon[50931]: Found osd claims for drivegroup all-available-devices -> {'vm01': ['1']} 2026-03-06T10:02:38.866 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:38 vm01 ceph-mon[50931]: from='client.14608 -' entity='client.admin' cmd=[{"prefix": "orch osd rm status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:38.866 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:38 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/3030623918' entity='client.admin' cmd=[{"prefix": "osd dump"}]: dispatch 2026-03-06T10:02:38.866 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:38 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/3938268074' entity='client.bootstrap-osd' cmd=[{"prefix": "osd tree", "format": "json"}]: dispatch 2026-03-06T10:02:39.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:38 vm07 ceph-mon[55141]: Detected new or changed devices on vm01 2026-03-06T10:02:39.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:38 vm07 ceph-mon[55141]: Found osd claims -> {'vm01': ['1']} 2026-03-06T10:02:39.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:38 vm07 ceph-mon[55141]: Found osd claims for drivegroup all-available-devices -> {'vm01': ['1']} 2026-03-06T10:02:39.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:38 vm07 ceph-mon[55141]: from='client.14608 -' entity='client.admin' cmd=[{"prefix": "orch osd rm status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T10:02:39.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:38 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/3030623918' entity='client.admin' cmd=[{"prefix": "osd dump"}]: dispatch 2026-03-06T10:02:39.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:38 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/3938268074' entity='client.bootstrap-osd' cmd=[{"prefix": "osd tree", "format": "json"}]: dispatch 2026-03-06T10:02:40.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:39 vm07 ceph-mon[55141]: pgmap v76: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail; 56 KiB/s, 0 objects/s recovering 2026-03-06T10:02:40.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:39 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/1431238015' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "fa0d5d34-5b91-4f08-9cec-a38a28e067cd", "id": 1}]: dispatch 2026-03-06T10:02:40.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:39 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/1431238015' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "fa0d5d34-5b91-4f08-9cec-a38a28e067cd", "id": 1}]': finished 2026-03-06T10:02:40.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:39 vm07 ceph-mon[55141]: osdmap e34: 8 total, 7 up, 7 in 2026-03-06T10:02:40.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:39 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/1551899555' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T10:02:40.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:39 vm01 ceph-mon[50931]: pgmap v76: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail; 56 KiB/s, 0 objects/s recovering 2026-03-06T10:02:40.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:39 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/1431238015' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "fa0d5d34-5b91-4f08-9cec-a38a28e067cd", "id": 1}]: dispatch 2026-03-06T10:02:40.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:39 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/1431238015' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "fa0d5d34-5b91-4f08-9cec-a38a28e067cd", "id": 1}]': finished 2026-03-06T10:02:40.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:39 vm01 ceph-mon[50931]: osdmap e34: 8 total, 7 up, 7 in 2026-03-06T10:02:40.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:39 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/1551899555' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T10:02:42.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:41 vm07 ceph-mon[55141]: pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-06T10:02:42.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:41 vm01 ceph-mon[50931]: pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-06T10:02:43.221 INFO:teuthology.orchestra.run.vm01.stderr:+ ceph osd dump 2026-03-06T10:02:43.225 INFO:teuthology.orchestra.run.vm01.stderr:+ grep 'up\s*in' 2026-03-06T10:02:43.226 INFO:teuthology.orchestra.run.vm01.stderr:+ grep osd.1 2026-03-06T10:02:43.401 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-06T10:02:44.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:43 vm07 ceph-mon[55141]: pgmap v79: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-06T10:02:44.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:43 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/4003977845' entity='client.admin' cmd=[{"prefix": "osd dump"}]: dispatch 2026-03-06T10:02:44.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:43 vm01 ceph-mon[50931]: pgmap v79: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-06T10:02:44.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:43 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/4003977845' entity='client.admin' cmd=[{"prefix": "osd dump"}]: dispatch 2026-03-06T10:02:45.165 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:44 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-06T10:02:45.165 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:44 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:02:45.165 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:44 vm01 ceph-mon[50931]: Deploying daemon osd.1 on vm01 2026-03-06T10:02:45.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:44 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-06T10:02:45.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:44 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:02:45.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:44 vm07 ceph-mon[55141]: Deploying daemon osd.1 on vm01 2026-03-06T10:02:46.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:45 vm07 ceph-mon[55141]: pgmap v80: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-06T10:02:46.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:45 vm01 ceph-mon[50931]: pgmap v80: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-06T10:02:47.931 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:47 vm01 ceph-mon[50931]: pgmap v81: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-06T10:02:47.931 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:47 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:47.931 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:47 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:48.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:47 vm07 ceph-mon[55141]: pgmap v81: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-06T10:02:48.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:47 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:48.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:47 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:48.403 INFO:teuthology.orchestra.run.vm01.stderr:+ grep osd.1 2026-03-06T10:02:48.403 INFO:teuthology.orchestra.run.vm01.stderr:+ grep 'up\s*in' 2026-03-06T10:02:48.403 INFO:teuthology.orchestra.run.vm01.stderr:+ ceph osd dump 2026-03-06T10:02:48.709 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 5 2026-03-06T10:02:49.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:48 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/2890433306' entity='client.admin' cmd=[{"prefix": "osd dump"}]: dispatch 2026-03-06T10:02:49.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:48 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T10:02:49.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:48 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/2890433306' entity='client.admin' cmd=[{"prefix": "osd dump"}]: dispatch 2026-03-06T10:02:49.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:48 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T10:02:50.043 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:49 vm07 ceph-mon[55141]: pgmap v82: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-06T10:02:50.043 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:49 vm07 ceph-mon[55141]: from='osd.1 [v2:192.168.123.101:6802/2776041583,v1:192.168.123.101:6803/2776041583]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-06T10:02:50.043 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:49 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:50.043 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:49 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:50.043 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:49 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T10:02:50.127 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:49 vm01 ceph-mon[50931]: pgmap v82: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-06T10:02:50.127 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:49 vm01 ceph-mon[50931]: from='osd.1 [v2:192.168.123.101:6802/2776041583,v1:192.168.123.101:6803/2776041583]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-06T10:02:50.128 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:49 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:50.128 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:49 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:50.128 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:49 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T10:02:51.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:50 vm01 ceph-mon[50931]: from='osd.1 [v2:192.168.123.101:6802/2776041583,v1:192.168.123.101:6803/2776041583]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-06T10:02:51.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:50 vm01 ceph-mon[50931]: osdmap e35: 8 total, 7 up, 7 in 2026-03-06T10:02:51.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:50 vm01 ceph-mon[50931]: from='osd.1 [v2:192.168.123.101:6802/2776041583,v1:192.168.123.101:6803/2776041583]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-06T10:02:51.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:50 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T10:02:51.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:50 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:51.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:50 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:51.018 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:50 vm01 ceph-mon[50931]: from='osd.1 [v2:192.168.123.101:6802/2776041583,v1:192.168.123.101:6803/2776041583]' entity='osd.1' 2026-03-06T10:02:51.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:50 vm07 ceph-mon[55141]: from='osd.1 [v2:192.168.123.101:6802/2776041583,v1:192.168.123.101:6803/2776041583]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-06T10:02:51.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:50 vm07 ceph-mon[55141]: osdmap e35: 8 total, 7 up, 7 in 2026-03-06T10:02:51.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:50 vm07 ceph-mon[55141]: from='osd.1 [v2:192.168.123.101:6802/2776041583,v1:192.168.123.101:6803/2776041583]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-06T10:02:51.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:50 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T10:02:51.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:50 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:51.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:50 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:51.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:50 vm07 ceph-mon[55141]: from='osd.1 [v2:192.168.123.101:6802/2776041583,v1:192.168.123.101:6803/2776041583]' entity='osd.1' 2026-03-06T10:02:52.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:51 vm07 ceph-mon[55141]: pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-06T10:02:52.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:51 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:52.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:51 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:52.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:51 vm07 ceph-mon[55141]: osd.1 [v2:192.168.123.101:6802/2776041583,v1:192.168.123.101:6803/2776041583] boot 2026-03-06T10:02:52.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:51 vm07 ceph-mon[55141]: osdmap e36: 8 total, 8 up, 8 in 2026-03-06T10:02:52.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:51 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T10:02:52.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:51 vm01 ceph-mon[50931]: pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-06T10:02:52.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:51 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:52.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:51 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:52.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:51 vm01 ceph-mon[50931]: osd.1 [v2:192.168.123.101:6802/2776041583,v1:192.168.123.101:6803/2776041583] boot 2026-03-06T10:02:52.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:51 vm01 ceph-mon[50931]: osdmap e36: 8 total, 8 up, 8 in 2026-03-06T10:02:52.268 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:51 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T10:02:53.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:52 vm07 ceph-mon[55141]: purged_snaps scrub starts 2026-03-06T10:02:53.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:52 vm07 ceph-mon[55141]: purged_snaps scrub ok 2026-03-06T10:02:53.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:52 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:53.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:52 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:53.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:52 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-06T10:02:53.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:52 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:02:53.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:52 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T10:02:53.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:52 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:53.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:52 vm07 ceph-mon[55141]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T10:02:53.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:52 vm07 ceph-mon[55141]: osdmap e37: 8 total, 8 up, 8 in 2026-03-06T10:02:53.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:52 vm01 ceph-mon[50931]: purged_snaps scrub starts 2026-03-06T10:02:53.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:52 vm01 ceph-mon[50931]: purged_snaps scrub ok 2026-03-06T10:02:53.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:52 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:53.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:52 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:53.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:52 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-06T10:02:53.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:52 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T10:02:53.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:52 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T10:02:53.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:52 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' 2026-03-06T10:02:53.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:52 vm01 ceph-mon[50931]: from='mgr.14219 192.168.123.101:0/2045785731' entity='mgr.vm01.ifwqbh' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T10:02:53.267 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:52 vm01 ceph-mon[50931]: osdmap e37: 8 total, 8 up, 8 in 2026-03-06T10:02:53.713 INFO:teuthology.orchestra.run.vm01.stderr:+ ceph osd dump 2026-03-06T10:02:53.713 INFO:teuthology.orchestra.run.vm01.stderr:+ grep 'up\s*in' 2026-03-06T10:02:53.713 INFO:teuthology.orchestra.run.vm01.stderr:+ grep osd.1 2026-03-06T10:02:53.878 INFO:teuthology.orchestra.run.vm01.stdout:osd.1 up in weight 1 up_from 36 up_thru 36 down_at 32 last_clean_interval [0,0) [v2:192.168.123.101:6802/2776041583,v1:192.168.123.101:6803/2776041583] [v2:192.168.123.101:6804/2776041583,v1:192.168.123.101:6805/2776041583] exists,up fa0d5d34-5b91-4f08-9cec-a38a28e067cd 2026-03-06T10:02:54.041 DEBUG:teuthology.run_tasks:Unwinding manager cephadm 2026-03-06T10:02:54.043 INFO:tasks.cephadm:Teardown begin 2026-03-06T10:02:54.043 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-06T10:02:54.068 DEBUG:teuthology.orchestra.run.vm07:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-06T10:02:54.092 INFO:tasks.cephadm:Cleaning up testdir ceph.* files... 2026-03-06T10:02:54.092 DEBUG:teuthology.orchestra.run.vm01:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-06T10:02:54.122 DEBUG:teuthology.orchestra.run.vm07:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-06T10:02:54.147 INFO:tasks.cephadm:Stopping all daemons... 2026-03-06T10:02:54.147 INFO:tasks.cephadm.mon.vm01:Stopping mon.vm01... 2026-03-06T10:02:54.147 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ceph-a3ab6672-193a-11f1-b81f-a119763c7190@mon.vm01 2026-03-06T10:02:54.183 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:53 vm01 ceph-mon[50931]: Detected new or changed devices on vm01 2026-03-06T10:02:54.183 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:53 vm01 ceph-mon[50931]: pgmap v87: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-06T10:02:54.183 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:53 vm01 ceph-mon[50931]: from='client.? 192.168.123.101:0/2758051595' entity='client.admin' cmd=[{"prefix": "osd dump"}]: dispatch 2026-03-06T10:02:54.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:53 vm07 ceph-mon[55141]: Detected new or changed devices on vm01 2026-03-06T10:02:54.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:53 vm07 ceph-mon[55141]: pgmap v87: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-06T10:02:54.196 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:53 vm07 ceph-mon[55141]: from='client.? 192.168.123.101:0/2758051595' entity='client.admin' cmd=[{"prefix": "osd dump"}]: dispatch 2026-03-06T10:02:54.468 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:54 vm01 systemd[1]: Stopping Ceph mon.vm01 for a3ab6672-193a-11f1-b81f-a119763c7190... 2026-03-06T10:02:54.469 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:54 vm01 ceph-a3ab6672-193a-11f1-b81f-a119763c7190-mon-vm01[50907]: 2026-03-06T09:02:54.264+0000 7fba06332640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.vm01 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-06T10:02:54.469 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:54 vm01 ceph-a3ab6672-193a-11f1-b81f-a119763c7190-mon-vm01[50907]: 2026-03-06T09:02:54.264+0000 7fba06332640 -1 mon.vm01@0(leader) e2 *** Got Signal Terminated *** 2026-03-06T10:02:54.698 DEBUG:teuthology.orchestra.run.vm01:> sudo pkill -f 'journalctl -f -n 0 -u ceph-a3ab6672-193a-11f1-b81f-a119763c7190@mon.vm01.service' 2026-03-06T10:02:54.766 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:54 vm01 podman[100163]: 2026-03-06 10:02:54.475738534 +0100 CET m=+0.227117887 container died f5dcc017bf8117f2a64e2f487c70d7a21b4a02231c940cd98f5208a6c9dd96e6 (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-a3ab6672-193a-11f1-b81f-a119763c7190-mon-vm01, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9) 2026-03-06T10:02:54.766 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:54 vm01 podman[100163]: 2026-03-06 10:02:54.604950656 +0100 CET m=+0.356330009 container remove f5dcc017bf8117f2a64e2f487c70d7a21b4a02231c940cd98f5208a6c9dd96e6 (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-a3ab6672-193a-11f1-b81f-a119763c7190-mon-vm01, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True) 2026-03-06T10:02:54.766 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:54 vm01 bash[100163]: ceph-a3ab6672-193a-11f1-b81f-a119763c7190-mon-vm01 2026-03-06T10:02:54.766 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:54 vm01 systemd[1]: ceph-a3ab6672-193a-11f1-b81f-a119763c7190@mon.vm01.service: Deactivated successfully. 2026-03-06T10:02:54.766 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:54 vm01 systemd[1]: Stopped Ceph mon.vm01 for a3ab6672-193a-11f1-b81f-a119763c7190. 2026-03-06T10:02:54.766 INFO:journalctl@ceph.mon.vm01.vm01.stdout:Mar 06 10:02:54 vm01 systemd[1]: ceph-a3ab6672-193a-11f1-b81f-a119763c7190@mon.vm01.service: Consumed 3.576s CPU time. 2026-03-06T10:02:54.783 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-06T10:02:54.783 INFO:tasks.cephadm.mon.vm01:Stopped mon.vm01 2026-03-06T10:02:54.783 INFO:tasks.cephadm.mon.vm07:Stopping mon.vm07... 2026-03-06T10:02:54.783 DEBUG:teuthology.orchestra.run.vm07:> sudo systemctl stop ceph-a3ab6672-193a-11f1-b81f-a119763c7190@mon.vm07 2026-03-06T10:02:54.963 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:54 vm07 systemd[1]: Stopping Ceph mon.vm07 for a3ab6672-193a-11f1-b81f-a119763c7190... 2026-03-06T10:02:54.963 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:54 vm07 ceph-a3ab6672-193a-11f1-b81f-a119763c7190-mon-vm07[55117]: 2026-03-06T09:02:54.903+0000 7f16d34a5640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.vm07 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-06T10:02:54.963 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:54 vm07 ceph-a3ab6672-193a-11f1-b81f-a119763c7190-mon-vm07[55117]: 2026-03-06T09:02:54.903+0000 7f16d34a5640 -1 mon.vm07@1(peon) e2 *** Got Signal Terminated *** 2026-03-06T10:02:55.274 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:55 vm07 podman[72507]: 2026-03-06 10:02:55.063539903 +0100 CET m=+0.174356396 container died e7653b40e4846e08aa290a8ba6753c6dc14e2049e81a2ea258034a4f16c3a672 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-a3ab6672-193a-11f1-b81f-a119763c7190-mon-vm07, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git) 2026-03-06T10:02:55.274 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:55 vm07 podman[72507]: 2026-03-06 10:02:55.183736766 +0100 CET m=+0.294553259 container remove e7653b40e4846e08aa290a8ba6753c6dc14e2049e81a2ea258034a4f16c3a672 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-a3ab6672-193a-11f1-b81f-a119763c7190-mon-vm07, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True) 2026-03-06T10:02:55.274 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 06 10:02:55 vm07 bash[72507]: ceph-a3ab6672-193a-11f1-b81f-a119763c7190-mon-vm07 2026-03-06T10:02:55.288 DEBUG:teuthology.orchestra.run.vm07:> sudo pkill -f 'journalctl -f -n 0 -u ceph-a3ab6672-193a-11f1-b81f-a119763c7190@mon.vm07.service' 2026-03-06T10:02:55.338 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-06T10:02:55.338 INFO:tasks.cephadm.mon.vm07:Stopped mon.vm07 2026-03-06T10:02:55.338 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid a3ab6672-193a-11f1-b81f-a119763c7190 --force --keep-logs 2026-03-06T10:02:55.712 INFO:teuthology.orchestra.run.vm01.stdout:Deleting cluster with fsid: a3ab6672-193a-11f1-b81f-a119763c7190 2026-03-06T10:03:24.238 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid a3ab6672-193a-11f1-b81f-a119763c7190 --force --keep-logs 2026-03-06T10:03:24.514 INFO:teuthology.orchestra.run.vm07.stdout:Deleting cluster with fsid: a3ab6672-193a-11f1-b81f-a119763c7190 2026-03-06T10:03:51.222 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-06T10:03:51.250 DEBUG:teuthology.orchestra.run.vm07:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-06T10:03:51.279 INFO:tasks.cephadm:Archiving crash dumps... 2026-03-06T10:03:51.279 DEBUG:teuthology.misc:Transferring archived files from vm01:/var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/crash to /archive/irq0-2026-03-06_09:41:01-orch:cephadm:osds-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/115/remote/vm01/crash 2026-03-06T10:03:51.279 DEBUG:teuthology.orchestra.run.vm01:> sudo tar c -f - -C /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/crash -- . 2026-03-06T10:03:51.318 INFO:teuthology.orchestra.run.vm01.stderr:tar: /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/crash: Cannot open: No such file or directory 2026-03-06T10:03:51.318 INFO:teuthology.orchestra.run.vm01.stderr:tar: Error is not recoverable: exiting now 2026-03-06T10:03:51.319 DEBUG:teuthology.misc:Transferring archived files from vm07:/var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/crash to /archive/irq0-2026-03-06_09:41:01-orch:cephadm:osds-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/115/remote/vm07/crash 2026-03-06T10:03:51.319 DEBUG:teuthology.orchestra.run.vm07:> sudo tar c -f - -C /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/crash -- . 2026-03-06T10:03:51.345 INFO:teuthology.orchestra.run.vm07.stderr:tar: /var/lib/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/crash: Cannot open: No such file or directory 2026-03-06T10:03:51.345 INFO:teuthology.orchestra.run.vm07.stderr:tar: Error is not recoverable: exiting now 2026-03-06T10:03:51.346 INFO:tasks.cephadm:Checking cluster log for badness... 2026-03-06T10:03:51.346 DEBUG:teuthology.orchestra.run.vm01:> sudo egrep '\[ERR\]|\[WRN\]|\[SEC\]' /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph.log | egrep CEPHADM_ | egrep -v '\(MDS_ALL_DOWN\)' | egrep -v '\(MDS_UP_LESS_THAN_MAX\)' | egrep -v OSD_DOWN | egrep -v CEPHADM_FAILED_DAEMON | egrep -v 'but is still running' | egrep -v PG_DEGRADED | head -n 1 2026-03-06T10:03:51.391 INFO:tasks.cephadm:Compressing logs... 2026-03-06T10:03:51.392 DEBUG:teuthology.orchestra.run.vm01:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-06T10:03:51.433 DEBUG:teuthology.orchestra.run.vm07:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-06T10:03:51.458 INFO:teuthology.orchestra.run.vm07.stderr:find: gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-06T10:03:51.459 INFO:teuthology.orchestra.run.vm07.stderr:‘/var/log/rbd-target-api’: No such file or directory 2026-03-06T10:03:51.459 INFO:teuthology.orchestra.run.vm07.stderr:gzip -5 --verbose -- /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-volume.log 2026-03-06T10:03:51.460 INFO:teuthology.orchestra.run.vm07.stderr:/var/log/ceph/cephadm.log: 92.1% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-06T10:03:51.461 INFO:teuthology.orchestra.run.vm07.stderr:gzip -5 --verbose -- /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-client.ceph-exporter.vm07.log 2026-03-06T10:03:51.461 INFO:teuthology.orchestra.run.vm07.stderr:gzip -5 --verbose -- /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-mgr.vm07.myglqt.log 2026-03-06T10:03:51.461 INFO:teuthology.orchestra.run.vm07.stderr:/var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-volume.log: /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-client.ceph-exporter.vm07.log: 29.6% -- replaced with /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-client.ceph-exporter.vm07.log.gz 2026-03-06T10:03:51.462 INFO:teuthology.orchestra.run.vm01.stderr:find: gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-06T10:03:51.462 INFO:teuthology.orchestra.run.vm01.stderr:‘/var/log/rbd-target-api’: No such file or directory 2026-03-06T10:03:51.462 INFO:teuthology.orchestra.run.vm07.stderr:gzip -5 --verbose -- /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-mon.vm07.log 2026-03-06T10:03:51.462 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-mon.vm01.log 2026-03-06T10:03:51.463 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/cephadm.log: gzip -5 --verbose -- /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph.log 2026-03-06T10:03:51.465 INFO:teuthology.orchestra.run.vm07.stderr:/var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-mgr.vm07.myglqt.log: 90.8% -- replaced with /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-mgr.vm07.myglqt.log.gz 2026-03-06T10:03:51.465 INFO:teuthology.orchestra.run.vm07.stderr:gzip -5 --verbose -- /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph.audit.log 2026-03-06T10:03:51.470 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-mon.vm01.log: gzip -5 --verbose -- /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph.audit.log 2026-03-06T10:03:51.470 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph.log: 84.5% -- replaced with /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph.log.gz 2026-03-06T10:03:51.470 INFO:teuthology.orchestra.run.vm07.stderr:/var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-mon.vm07.log: gzip -5 --verbose -- /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph.log 2026-03-06T10:03:51.473 INFO:teuthology.orchestra.run.vm07.stderr:/var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph.audit.log: 90.9% -- replaced with /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph.audit.log.gz 2026-03-06T10:03:51.475 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-mgr.vm01.ifwqbh.log 2026-03-06T10:03:51.475 INFO:teuthology.orchestra.run.vm07.stderr:gzip -5 --verbose -- /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph.cephadm.log 2026-03-06T10:03:51.475 INFO:teuthology.orchestra.run.vm07.stderr: 93.4% -- replaced with /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-volume.log.gz 2026-03-06T10:03:51.475 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph.audit.log: 92.0% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-06T10:03:51.476 INFO:teuthology.orchestra.run.vm07.stderr:/var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph.log: 83.6% -- replaced with /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph.log.gz 2026-03-06T10:03:51.476 INFO:teuthology.orchestra.run.vm01.stderr: 90.7% -- replaced with /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph.audit.log.gz 2026-03-06T10:03:51.476 INFO:teuthology.orchestra.run.vm07.stderr:gzip -5 --verbose -- /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-osd.0.log 2026-03-06T10:03:51.476 INFO:teuthology.orchestra.run.vm07.stderr:/var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph.cephadm.log: 81.2% -- replaced with /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph.cephadm.log.gz 2026-03-06T10:03:51.477 INFO:teuthology.orchestra.run.vm07.stderr:gzip -5 --verbose -- /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-osd.3.log 2026-03-06T10:03:51.481 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph.cephadm.log 2026-03-06T10:03:51.482 INFO:teuthology.orchestra.run.vm07.stderr:/var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-osd.0.log: gzip -5 --verbose -- /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-osd.5.log 2026-03-06T10:03:51.486 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-mgr.vm01.ifwqbh.log: gzip -5 --verbose -- /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-volume.log 2026-03-06T10:03:51.486 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph.cephadm.log: 82.4% -- replaced with /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph.cephadm.log.gz 2026-03-06T10:03:51.490 INFO:teuthology.orchestra.run.vm07.stderr:/var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-osd.3.log: 92.1% -- replaced with /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-mon.vm07.log.gz 2026-03-06T10:03:51.491 INFO:teuthology.orchestra.run.vm07.stderr:gzip -5 --verbose -- /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-osd.7.log 2026-03-06T10:03:51.498 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-client.ceph-exporter.vm01.log 2026-03-06T10:03:51.509 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-volume.log: gzip -5 --verbose -- /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-osd.1.log 2026-03-06T10:03:51.509 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-client.ceph-exporter.vm01.log: 92.1% -- replaced with /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-client.ceph-exporter.vm01.log.gz 2026-03-06T10:03:51.512 INFO:teuthology.orchestra.run.vm07.stderr:/var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-osd.5.log: /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-osd.7.log: 92.7% -- replaced with /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-osd.3.log.gz 2026-03-06T10:03:51.512 INFO:teuthology.orchestra.run.vm07.stderr: 92.8% -- replaced with /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-osd.0.log.gz 2026-03-06T10:03:51.513 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-osd.2.log 2026-03-06T10:03:51.516 INFO:teuthology.orchestra.run.vm07.stderr: 92.7% -- replaced with /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-osd.5.log.gz 2026-03-06T10:03:51.521 INFO:teuthology.orchestra.run.vm07.stderr: 92.7% -- replaced with /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-osd.7.log.gz 2026-03-06T10:03:51.522 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-osd.1.log: gzip -5 --verbose -- /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-osd.4.log 2026-03-06T10:03:51.522 INFO:teuthology.orchestra.run.vm07.stderr: 2026-03-06T10:03:51.522 INFO:teuthology.orchestra.run.vm07.stderr:real 0m0.073s 2026-03-06T10:03:51.522 INFO:teuthology.orchestra.run.vm07.stderr:user 0m0.102s 2026-03-06T10:03:51.522 INFO:teuthology.orchestra.run.vm07.stderr:sys 0m0.028s 2026-03-06T10:03:51.535 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-osd.2.log: gzip -5 --verbose -- /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-osd.6.log 2026-03-06T10:03:51.549 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-osd.4.log: /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-osd.6.log: 93.1% -- replaced with /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-volume.log.gz 2026-03-06T10:03:51.574 INFO:teuthology.orchestra.run.vm01.stderr: 92.6% -- replaced with /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-osd.4.log.gz 2026-03-06T10:03:51.575 INFO:teuthology.orchestra.run.vm01.stderr: 89.6% -- replaced with /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-mgr.vm01.ifwqbh.log.gz 2026-03-06T10:03:51.576 INFO:teuthology.orchestra.run.vm01.stderr: 92.6% -- replaced with /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-osd.2.log.gz 2026-03-06T10:03:51.579 INFO:teuthology.orchestra.run.vm01.stderr: 93.3% -- replaced with /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-osd.1.log.gz 2026-03-06T10:03:51.579 INFO:teuthology.orchestra.run.vm01.stderr: 93.0% -- replaced with /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-osd.6.log.gz 2026-03-06T10:03:51.617 INFO:teuthology.orchestra.run.vm01.stderr: 91.4% -- replaced with /var/log/ceph/a3ab6672-193a-11f1-b81f-a119763c7190/ceph-mon.vm01.log.gz 2026-03-06T10:03:51.618 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-06T10:03:51.618 INFO:teuthology.orchestra.run.vm01.stderr:real 0m0.166s 2026-03-06T10:03:51.619 INFO:teuthology.orchestra.run.vm01.stderr:user 0m0.252s 2026-03-06T10:03:51.619 INFO:teuthology.orchestra.run.vm01.stderr:sys 0m0.029s 2026-03-06T10:03:51.619 INFO:tasks.cephadm:Archiving logs... 2026-03-06T10:03:51.619 DEBUG:teuthology.misc:Transferring archived files from vm01:/var/log/ceph to /archive/irq0-2026-03-06_09:41:01-orch:cephadm:osds-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/115/remote/vm01/log 2026-03-06T10:03:51.619 DEBUG:teuthology.orchestra.run.vm01:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-06T10:03:51.705 DEBUG:teuthology.misc:Transferring archived files from vm07:/var/log/ceph to /archive/irq0-2026-03-06_09:41:01-orch:cephadm:osds-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/115/remote/vm07/log 2026-03-06T10:03:51.705 DEBUG:teuthology.orchestra.run.vm07:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-06T10:03:51.738 INFO:tasks.cephadm:Removing cluster... 2026-03-06T10:03:51.739 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid a3ab6672-193a-11f1-b81f-a119763c7190 --force 2026-03-06T10:03:52.021 INFO:teuthology.orchestra.run.vm01.stdout:Deleting cluster with fsid: a3ab6672-193a-11f1-b81f-a119763c7190 2026-03-06T10:03:52.116 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid a3ab6672-193a-11f1-b81f-a119763c7190 --force 2026-03-06T10:03:52.397 INFO:teuthology.orchestra.run.vm07.stdout:Deleting cluster with fsid: a3ab6672-193a-11f1-b81f-a119763c7190 2026-03-06T10:03:52.494 INFO:tasks.cephadm:Removing cephadm ... 2026-03-06T10:03:52.494 DEBUG:teuthology.orchestra.run.vm01:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-06T10:03:52.509 DEBUG:teuthology.orchestra.run.vm07:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-06T10:03:52.524 INFO:tasks.cephadm:Teardown complete 2026-03-06T10:03:52.524 DEBUG:teuthology.run_tasks:Unwinding manager clock 2026-03-06T10:03:52.527 INFO:teuthology.task.clock:Checking final clock skew... 2026-03-06T10:03:52.527 DEBUG:teuthology.orchestra.run.vm01:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-06T10:03:52.551 DEBUG:teuthology.orchestra.run.vm07:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-06T10:03:52.564 INFO:teuthology.orchestra.run.vm01.stderr:bash: line 1: ntpq: command not found 2026-03-06T10:03:52.567 INFO:teuthology.orchestra.run.vm01.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-06T10:03:52.567 INFO:teuthology.orchestra.run.vm01.stdout:=============================================================================== 2026-03-06T10:03:52.568 INFO:teuthology.orchestra.run.vm01.stdout:^* 158.101.188.125 2 6 377 2 +242us[ +224us] +/- 13ms 2026-03-06T10:03:52.568 INFO:teuthology.orchestra.run.vm01.stdout:^+ red-pelican-63749.zap.cl> 3 6 377 1 -688us[ -688us] +/- 32ms 2026-03-06T10:03:52.568 INFO:teuthology.orchestra.run.vm01.stdout:^- x1.ncomputers.org 2 6 377 65 +1092us[+1078us] +/- 50ms 2026-03-06T10:03:52.568 INFO:teuthology.orchestra.run.vm01.stdout:^+ listserver.trexler.at 2 6 377 64 +49us[ +36us] +/- 20ms 2026-03-06T10:03:52.580 INFO:teuthology.orchestra.run.vm07.stderr:bash: line 1: ntpq: command not found 2026-03-06T10:03:52.583 INFO:teuthology.orchestra.run.vm07.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-06T10:03:52.583 INFO:teuthology.orchestra.run.vm07.stdout:=============================================================================== 2026-03-06T10:03:52.583 INFO:teuthology.orchestra.run.vm07.stdout:^+ listserver.trexler.at 2 6 377 1 +55us[ +55us] +/- 20ms 2026-03-06T10:03:52.583 INFO:teuthology.orchestra.run.vm07.stdout:^* 158.101.188.125 2 6 377 2 +242us[ +232us] +/- 13ms 2026-03-06T10:03:52.583 INFO:teuthology.orchestra.run.vm07.stdout:^+ red-pelican-63749.zap.cl> 3 6 377 0 -695us[ -695us] +/- 32ms 2026-03-06T10:03:52.583 INFO:teuthology.orchestra.run.vm07.stdout:^- x1.ncomputers.org 2 6 377 0 -149us[ -149us] +/- 49ms 2026-03-06T10:03:52.583 DEBUG:teuthology.run_tasks:Unwinding manager ansible.cephlab 2026-03-06T10:03:52.587 INFO:teuthology.task.ansible:Skipping ansible cleanup... 2026-03-06T10:03:52.587 DEBUG:teuthology.run_tasks:Unwinding manager selinux 2026-03-06T10:03:52.589 DEBUG:teuthology.run_tasks:Unwinding manager pcp 2026-03-06T10:03:52.591 DEBUG:teuthology.run_tasks:Unwinding manager internal.timer 2026-03-06T10:03:52.595 INFO:teuthology.task.internal:Duration was 461.871637 seconds 2026-03-06T10:03:52.596 DEBUG:teuthology.run_tasks:Unwinding manager internal.syslog 2026-03-06T10:03:52.597 INFO:teuthology.task.internal.syslog:Shutting down syslog monitoring... 2026-03-06T10:03:52.597 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-06T10:03:52.610 DEBUG:teuthology.orchestra.run.vm07:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-06T10:03:52.645 INFO:teuthology.orchestra.run.vm01.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-06T10:03:52.661 INFO:teuthology.orchestra.run.vm07.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-06T10:03:53.054 INFO:teuthology.task.internal.syslog:Checking logs for errors... 2026-03-06T10:03:53.054 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm01.local 2026-03-06T10:03:53.054 DEBUG:teuthology.orchestra.run.vm01:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-06T10:03:53.116 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm07.local 2026-03-06T10:03:53.116 DEBUG:teuthology.orchestra.run.vm07:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-06T10:03:53.139 INFO:teuthology.task.internal.syslog:Gathering journactl... 2026-03-06T10:03:53.140 DEBUG:teuthology.orchestra.run.vm01:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-06T10:03:53.158 DEBUG:teuthology.orchestra.run.vm07:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-06T10:03:53.646 INFO:teuthology.task.internal.syslog:Compressing syslogs... 2026-03-06T10:03:53.647 DEBUG:teuthology.orchestra.run.vm01:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-06T10:03:53.648 DEBUG:teuthology.orchestra.run.vm07:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-06T10:03:53.672 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-06T10:03:53.672 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-06T10:03:53.672 INFO:teuthology.orchestra.run.vm01.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-06T10:03:53.672 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-06T10:03:53.672 INFO:teuthology.orchestra.run.vm07.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-06T10:03:53.672 INFO:teuthology.orchestra.run.vm07.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-06T10:03:53.673 INFO:teuthology.orchestra.run.vm07.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-06T10:03:53.673 INFO:teuthology.orchestra.run.vm01.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-06T10:03:53.673 INFO:teuthology.orchestra.run.vm07.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-06T10:03:53.673 INFO:teuthology.orchestra.run.vm07.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: /home/ubuntu/cephtest/archive/syslog/journalctl.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-06T10:03:53.789 INFO:teuthology.orchestra.run.vm07.stderr: 98.3% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-06T10:03:53.823 INFO:teuthology.orchestra.run.vm01.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 98.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-06T10:03:53.825 DEBUG:teuthology.run_tasks:Unwinding manager internal.sudo 2026-03-06T10:03:53.827 INFO:teuthology.task.internal:Restoring /etc/sudoers... 2026-03-06T10:03:53.827 DEBUG:teuthology.orchestra.run.vm01:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-06T10:03:53.888 DEBUG:teuthology.orchestra.run.vm07:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-06T10:03:53.912 DEBUG:teuthology.run_tasks:Unwinding manager internal.coredump 2026-03-06T10:03:53.914 DEBUG:teuthology.orchestra.run.vm01:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-06T10:03:53.931 DEBUG:teuthology.orchestra.run.vm07:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-06T10:03:53.954 INFO:teuthology.orchestra.run.vm01.stdout:kernel.core_pattern = core 2026-03-06T10:03:53.975 INFO:teuthology.orchestra.run.vm07.stdout:kernel.core_pattern = core 2026-03-06T10:03:53.989 DEBUG:teuthology.orchestra.run.vm01:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-06T10:03:54.027 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T10:03:54.027 DEBUG:teuthology.orchestra.run.vm07:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-06T10:03:54.043 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T10:03:54.043 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive 2026-03-06T10:03:54.045 INFO:teuthology.task.internal:Transferring archived files... 2026-03-06T10:03:54.046 DEBUG:teuthology.misc:Transferring archived files from vm01:/home/ubuntu/cephtest/archive to /archive/irq0-2026-03-06_09:41:01-orch:cephadm:osds-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/115/remote/vm01 2026-03-06T10:03:54.046 DEBUG:teuthology.orchestra.run.vm01:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-06T10:03:54.097 DEBUG:teuthology.misc:Transferring archived files from vm07:/home/ubuntu/cephtest/archive to /archive/irq0-2026-03-06_09:41:01-orch:cephadm:osds-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/115/remote/vm07 2026-03-06T10:03:54.097 DEBUG:teuthology.orchestra.run.vm07:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-06T10:03:54.128 INFO:teuthology.task.internal:Removing archive directory... 2026-03-06T10:03:54.128 DEBUG:teuthology.orchestra.run.vm01:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-06T10:03:54.139 DEBUG:teuthology.orchestra.run.vm07:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-06T10:03:54.184 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive_upload 2026-03-06T10:03:54.187 INFO:teuthology.task.internal:Not uploading archives. 2026-03-06T10:03:54.187 DEBUG:teuthology.run_tasks:Unwinding manager internal.base 2026-03-06T10:03:54.189 INFO:teuthology.task.internal:Tidying up after the test... 2026-03-06T10:03:54.189 DEBUG:teuthology.orchestra.run.vm01:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-06T10:03:54.199 DEBUG:teuthology.orchestra.run.vm07:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-06T10:03:54.213 INFO:teuthology.orchestra.run.vm01.stdout: 8532016 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 6 10:03 /home/ubuntu/cephtest 2026-03-06T10:03:54.243 INFO:teuthology.orchestra.run.vm07.stdout: 8532146 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 6 10:03 /home/ubuntu/cephtest 2026-03-06T10:03:54.244 DEBUG:teuthology.run_tasks:Unwinding manager console_log 2026-03-06T10:03:54.249 INFO:teuthology.run:Summary data: description: orch:cephadm:osds/{0-distro/centos_9.stream_runc 1-start 2-ops/rm-zap-flag} duration: 461.8716366291046 owner: irq0 success: true 2026-03-06T10:03:54.249 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-06T10:03:54.268 INFO:teuthology.run:pass