2026-03-07T10:04:39.857 INFO:root:teuthology version: 1.2.4.dev6+g1c580df7a 2026-03-07T10:04:39.863 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-07T10:04:39.883 INFO:teuthology.run:Config: archive_path: /archive/irq0-2026-03-07_10:02:54-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps/9 branch: cobaltcore-storage-v19.2.3-fasttrack-5 description: orch:cephadm:workunits/{0-distro/ubuntu_22.04 agent/off mon_election/classic task/test_extra_daemon_features} email: null first_in_suite: false flavor: default job_id: '9' ktype: distro last_in_suite: false machine_type: vps name: irq0-2026-03-07_10:02:54-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps no_nested_subset: false os_type: ubuntu os_version: '22.04' overrides: admin_socket: branch: cobaltcore-storage-v19.2.3-fasttrack-5 ansible.cephlab: branch: main repo: https://github.com/kshtsk/ceph-cm-ansible.git skip_tags: nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs vars: timezone: UTC ceph: conf: global: mon election default strategy: 1 mgr: debug mgr: 20 debug ms: 1 mgr/cephadm/use_agent: false mon: debug mon: 20 debug ms: 1 debug paxos: 20 osd: debug ms: 1 debug osd: 20 osd mclock iops capacity threshold hdd: 49000 flavor: default log-ignorelist: - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - CEPHADM_FAILED_DAEMON log-only-match: - CEPHADM_ sha1: 340d3c24fc6ae7529322dc7ccee6c6cb2589da0a ceph-deploy: conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: {} cephadm: cephadm_binary_url: https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm containers: image: harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 install: ceph: flavor: default sha1: 340d3c24fc6ae7529322dc7ccee6c6cb2589da0a extra_system_packages: deb: - python3-xmltodict - s3cmd rpm: - bzip2 - perl-Test-Harness - python3-xmltodict - s3cmd repos: - name: ceph-source priority: 1 url: https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-39-g340d3c24fc6/el9.clyso/SRPMS - name: ceph-noarch priority: 1 url: https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-39-g340d3c24fc6/el9.clyso/noarch - name: ceph priority: 1 url: https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-39-g340d3c24fc6/el9.clyso/x86_64 workunit: branch: tt-fasttrack-5-workunits sha1: f96e33505a05da25eb24b46ae34fbbd1718a702b owner: irq0 priority: 1000 repo: https://github.com/ceph/ceph.git roles: - - host.a - mon.a - mgr.a - osd.0 - - host.b - mon.b - mgr.b - osd.1 seed: 8363 sha1: 340d3c24fc6ae7529322dc7ccee6c6cb2589da0a sleep_before_teardown: 0 subset: 1/64 suite: orch:cephadm:workunits suite_branch: tt-fasttrack-5-workunits suite_path: /home/teuthos/src/github.com_kshtsk_ceph_f96e33505a05da25eb24b46ae34fbbd1718a702b/qa suite_relpath: qa suite_repo: https://github.com/kshtsk/ceph.git suite_sha1: f96e33505a05da25eb24b46ae34fbbd1718a702b targets: vm02.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBJFlUiLulXJz6iVjZLhjFLXsDKjzm1eA675MAvXKP60MOk0aAfjj/66w6sn3utqssIBCGLAqR+CKHyATYVxKxt8= vm10.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBBkPHckySH7vSJaIF85A3EfDaf4KM0caW9RwUJNZk0zRv5Kksq7MZWmSLgmHwPA9GyAyiQt9ydlut1Q0iNmIERw= tasks: - cephadm: null - exec: all-hosts: - mkdir /etc/cephadm_testing - cephadm.apply: specs: - extra_container_args: - --cpus=2 extra_entrypoint_args: - --debug_ms 10 placement: host_pattern: '*' service_type: mon - custom_configs: - content: "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n\ \ ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\n\ sleep infinity\n" mount_path: /root/write_thing_to_file.sh extra_container_args: - -v - /etc/cephadm_testing:/root/cephadm_testing extra_entrypoint_args: - /root/write_thing_to_file.sh - -c - testing_custom_containers - -o - /root/cephadm_testing/testing.txt placement: host_pattern: '*' service_id: foo service_type: container spec: entrypoint: bash image: quay.io/fedora/fedora:latest - custom_configs: - content: 'set -e test -f /var/cache/bar/from.txt test -f /var/cache/bar/presized.dat echo ok > /var/cache/bar/primary.txt sleep infinity ' mount_path: /root/init_check.sh extra_entrypoint_args: - /root/init_check.sh placement: host_pattern: '*' service_id: bar service_type: container spec: dirs: - data entrypoint: bash image: quay.io/fedora/fedora:latest init_containers: - entrypoint: bash entrypoint_args: - argument: -c - argument: . /etc/os-release && echo from=$ID > /var/cache/bar/from.txt image: quay.io/centos/centos:latest volume_mounts: data: /var/cache/bar:z - entrypoint: bash entrypoint_args: - argument: -c - argument: test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat volume_mounts: data: /var/cache/bar:z volume_mounts: data: /var/cache/bar:z - cephadm.wait_for_service: service: mon - cephadm.wait_for_service: service: container.foo - cephadm.wait_for_service: service: container.bar - exec: host.a: - 'set -ex FSID=$(/home/ubuntu/cephtest/cephadm shell -- ceph fsid) sleep 60 # check extra container and entrypoint args written to mon unit run file grep "\-\-cpus=2" /var/lib/ceph/$FSID/mon.*/unit.run grep "\-\-debug_ms 10" /var/lib/ceph/$FSID/mon.*/unit.run # check that custom container properly wrote content to file. # This requires the custom config, extra container args, and # entrypoint args to all be working in order for this to have # been written. The container entrypoint was set up with custom_configs, # the content and where to write to with the entrypoint args, and the mounting # of the /etc/cephadm_testing dir with extra container args grep "testing_custom_containers" /etc/cephadm_testing/testing.txt # Verify that container bar''s init containers and primary container # ran successfully dir=$(find /var/lib/ceph/$FSID -maxdepth 1 -type d -name ''container.bar.*'') test -n "$dir" grep ok ${dir}/data/primary.txt grep from=centos ${dir}/data/from.txt test -s ${dir}/data/presized.dat ' teuthology: fragments_dropped: [] meta: {} postmerge: [] teuthology_branch: clyso-debian-13 teuthology_repo: https://github.com/clyso/teuthology teuthology_sha1: 1c580df7a9c7c2aadc272da296344fd99f27c444 timestamp: 2026-03-07_10:02:54 tube: vps user: irq0 verbose: false worker_log: /home/teuthos/.teuthology/dispatcher/dispatcher.vps.2764 2026-03-07T10:04:39.883 INFO:teuthology.run:suite_path is set to /home/teuthos/src/github.com_kshtsk_ceph_f96e33505a05da25eb24b46ae34fbbd1718a702b/qa; will attempt to use it 2026-03-07T10:04:39.884 INFO:teuthology.run:Found tasks at /home/teuthos/src/github.com_kshtsk_ceph_f96e33505a05da25eb24b46ae34fbbd1718a702b/qa/tasks 2026-03-07T10:04:39.884 INFO:teuthology.run_tasks:Running task internal.save_config... 2026-03-07T10:04:39.884 INFO:teuthology.task.internal:Saving configuration 2026-03-07T10:04:39.890 INFO:teuthology.run_tasks:Running task internal.check_lock... 2026-03-07T10:04:39.890 INFO:teuthology.task.internal.check_lock:Checking locks... 2026-03-07T10:04:39.895 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm02.local', 'description': '/archive/irq0-2026-03-07_10:02:54-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps/9', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'ubuntu', 'os_version': '22.04', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-07 10:03:36.784630', 'locked_by': 'irq0', 'mac_address': '52:55:00:00:00:02', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBJFlUiLulXJz6iVjZLhjFLXsDKjzm1eA675MAvXKP60MOk0aAfjj/66w6sn3utqssIBCGLAqR+CKHyATYVxKxt8='} 2026-03-07T10:04:39.900 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm10.local', 'description': '/archive/irq0-2026-03-07_10:02:54-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps/9', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'ubuntu', 'os_version': '22.04', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-07 10:03:36.785142', 'locked_by': 'irq0', 'mac_address': '52:55:00:00:00:0a', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBBkPHckySH7vSJaIF85A3EfDaf4KM0caW9RwUJNZk0zRv5Kksq7MZWmSLgmHwPA9GyAyiQt9ydlut1Q0iNmIERw='} 2026-03-07T10:04:39.900 INFO:teuthology.run_tasks:Running task internal.add_remotes... 2026-03-07T10:04:39.900 INFO:teuthology.task.internal:roles: ubuntu@vm02.local - ['host.a', 'mon.a', 'mgr.a', 'osd.0'] 2026-03-07T10:04:39.900 INFO:teuthology.task.internal:roles: ubuntu@vm10.local - ['host.b', 'mon.b', 'mgr.b', 'osd.1'] 2026-03-07T10:04:39.900 INFO:teuthology.run_tasks:Running task console_log... 2026-03-07T10:04:39.905 DEBUG:teuthology.task.console_log:vm02 does not support IPMI; excluding 2026-03-07T10:04:39.911 DEBUG:teuthology.task.console_log:vm10 does not support IPMI; excluding 2026-03-07T10:04:39.911 DEBUG:teuthology.exit:Installing handler: Handler(exiter=, func=.kill_console_loggers at 0x7f946bfd3be0>, signals=[15]) 2026-03-07T10:04:39.911 INFO:teuthology.run_tasks:Running task internal.connect... 2026-03-07T10:04:39.911 INFO:teuthology.task.internal:Opening connections... 2026-03-07T10:04:39.912 DEBUG:teuthology.task.internal:connecting to ubuntu@vm02.local 2026-03-07T10:04:39.912 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm02.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-07T10:04:39.971 DEBUG:teuthology.task.internal:connecting to ubuntu@vm10.local 2026-03-07T10:04:39.971 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm10.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-07T10:04:40.030 INFO:teuthology.run_tasks:Running task internal.push_inventory... 2026-03-07T10:04:40.031 DEBUG:teuthology.orchestra.run.vm02:> uname -m 2026-03-07T10:04:40.051 INFO:teuthology.orchestra.run.vm02.stdout:x86_64 2026-03-07T10:04:40.051 DEBUG:teuthology.orchestra.run.vm02:> cat /etc/os-release 2026-03-07T10:04:40.096 INFO:teuthology.orchestra.run.vm02.stdout:PRETTY_NAME="Ubuntu 22.04.5 LTS" 2026-03-07T10:04:40.096 INFO:teuthology.orchestra.run.vm02.stdout:NAME="Ubuntu" 2026-03-07T10:04:40.096 INFO:teuthology.orchestra.run.vm02.stdout:VERSION_ID="22.04" 2026-03-07T10:04:40.096 INFO:teuthology.orchestra.run.vm02.stdout:VERSION="22.04.5 LTS (Jammy Jellyfish)" 2026-03-07T10:04:40.096 INFO:teuthology.orchestra.run.vm02.stdout:VERSION_CODENAME=jammy 2026-03-07T10:04:40.096 INFO:teuthology.orchestra.run.vm02.stdout:ID=ubuntu 2026-03-07T10:04:40.096 INFO:teuthology.orchestra.run.vm02.stdout:ID_LIKE=debian 2026-03-07T10:04:40.096 INFO:teuthology.orchestra.run.vm02.stdout:HOME_URL="https://www.ubuntu.com/" 2026-03-07T10:04:40.096 INFO:teuthology.orchestra.run.vm02.stdout:SUPPORT_URL="https://help.ubuntu.com/" 2026-03-07T10:04:40.096 INFO:teuthology.orchestra.run.vm02.stdout:BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/" 2026-03-07T10:04:40.096 INFO:teuthology.orchestra.run.vm02.stdout:PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy" 2026-03-07T10:04:40.096 INFO:teuthology.orchestra.run.vm02.stdout:UBUNTU_CODENAME=jammy 2026-03-07T10:04:40.096 INFO:teuthology.lock.ops:Updating vm02.local on lock server 2026-03-07T10:04:40.100 DEBUG:teuthology.orchestra.run.vm10:> uname -m 2026-03-07T10:04:40.103 INFO:teuthology.orchestra.run.vm10.stdout:x86_64 2026-03-07T10:04:40.103 DEBUG:teuthology.orchestra.run.vm10:> cat /etc/os-release 2026-03-07T10:04:40.149 INFO:teuthology.orchestra.run.vm10.stdout:PRETTY_NAME="Ubuntu 22.04.5 LTS" 2026-03-07T10:04:40.149 INFO:teuthology.orchestra.run.vm10.stdout:NAME="Ubuntu" 2026-03-07T10:04:40.149 INFO:teuthology.orchestra.run.vm10.stdout:VERSION_ID="22.04" 2026-03-07T10:04:40.149 INFO:teuthology.orchestra.run.vm10.stdout:VERSION="22.04.5 LTS (Jammy Jellyfish)" 2026-03-07T10:04:40.149 INFO:teuthology.orchestra.run.vm10.stdout:VERSION_CODENAME=jammy 2026-03-07T10:04:40.149 INFO:teuthology.orchestra.run.vm10.stdout:ID=ubuntu 2026-03-07T10:04:40.149 INFO:teuthology.orchestra.run.vm10.stdout:ID_LIKE=debian 2026-03-07T10:04:40.149 INFO:teuthology.orchestra.run.vm10.stdout:HOME_URL="https://www.ubuntu.com/" 2026-03-07T10:04:40.149 INFO:teuthology.orchestra.run.vm10.stdout:SUPPORT_URL="https://help.ubuntu.com/" 2026-03-07T10:04:40.149 INFO:teuthology.orchestra.run.vm10.stdout:BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/" 2026-03-07T10:04:40.149 INFO:teuthology.orchestra.run.vm10.stdout:PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy" 2026-03-07T10:04:40.149 INFO:teuthology.orchestra.run.vm10.stdout:UBUNTU_CODENAME=jammy 2026-03-07T10:04:40.149 INFO:teuthology.lock.ops:Updating vm10.local on lock server 2026-03-07T10:04:40.154 INFO:teuthology.run_tasks:Running task internal.serialize_remote_roles... 2026-03-07T10:04:40.156 INFO:teuthology.run_tasks:Running task internal.check_conflict... 2026-03-07T10:04:40.157 INFO:teuthology.task.internal:Checking for old test directory... 2026-03-07T10:04:40.157 DEBUG:teuthology.orchestra.run.vm02:> test '!' -e /home/ubuntu/cephtest 2026-03-07T10:04:40.158 DEBUG:teuthology.orchestra.run.vm10:> test '!' -e /home/ubuntu/cephtest 2026-03-07T10:04:40.193 INFO:teuthology.run_tasks:Running task internal.check_ceph_data... 2026-03-07T10:04:40.194 INFO:teuthology.task.internal:Checking for non-empty /var/lib/ceph... 2026-03-07T10:04:40.194 DEBUG:teuthology.orchestra.run.vm02:> test -z $(ls -A /var/lib/ceph) 2026-03-07T10:04:40.202 DEBUG:teuthology.orchestra.run.vm10:> test -z $(ls -A /var/lib/ceph) 2026-03-07T10:04:40.204 INFO:teuthology.orchestra.run.vm02.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-07T10:04:40.237 INFO:teuthology.orchestra.run.vm10.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-07T10:04:40.238 INFO:teuthology.run_tasks:Running task internal.vm_setup... 2026-03-07T10:04:40.246 DEBUG:teuthology.orchestra.run.vm02:> test -e /ceph-qa-ready 2026-03-07T10:04:40.248 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-07T10:04:40.691 DEBUG:teuthology.orchestra.run.vm10:> test -e /ceph-qa-ready 2026-03-07T10:04:40.693 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-07T10:04:40.907 INFO:teuthology.run_tasks:Running task internal.base... 2026-03-07T10:04:40.908 INFO:teuthology.task.internal:Creating test directory... 2026-03-07T10:04:40.908 DEBUG:teuthology.orchestra.run.vm02:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-07T10:04:40.910 DEBUG:teuthology.orchestra.run.vm10:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-07T10:04:40.913 INFO:teuthology.run_tasks:Running task internal.archive_upload... 2026-03-07T10:04:40.913 INFO:teuthology.run_tasks:Running task internal.archive... 2026-03-07T10:04:40.915 INFO:teuthology.task.internal:Creating archive directory... 2026-03-07T10:04:40.915 DEBUG:teuthology.orchestra.run.vm02:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-07T10:04:40.954 DEBUG:teuthology.orchestra.run.vm10:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-07T10:04:40.958 INFO:teuthology.run_tasks:Running task internal.coredump... 2026-03-07T10:04:40.959 INFO:teuthology.task.internal:Enabling coredump saving... 2026-03-07T10:04:40.959 DEBUG:teuthology.orchestra.run.vm02:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-07T10:04:40.999 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-07T10:04:40.999 DEBUG:teuthology.orchestra.run.vm10:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-07T10:04:41.001 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-07T10:04:41.001 DEBUG:teuthology.orchestra.run.vm02:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-07T10:04:41.042 DEBUG:teuthology.orchestra.run.vm10:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-07T10:04:41.049 INFO:teuthology.orchestra.run.vm02.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-07T10:04:41.050 INFO:teuthology.orchestra.run.vm10.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-07T10:04:41.053 INFO:teuthology.orchestra.run.vm02.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-07T10:04:41.055 INFO:teuthology.orchestra.run.vm10.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-07T10:04:41.055 INFO:teuthology.run_tasks:Running task internal.sudo... 2026-03-07T10:04:41.057 INFO:teuthology.task.internal:Configuring sudo... 2026-03-07T10:04:41.057 DEBUG:teuthology.orchestra.run.vm02:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-07T10:04:41.098 DEBUG:teuthology.orchestra.run.vm10:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-07T10:04:41.105 INFO:teuthology.run_tasks:Running task internal.syslog... 2026-03-07T10:04:41.107 INFO:teuthology.task.internal.syslog:Starting syslog monitoring... 2026-03-07T10:04:41.107 DEBUG:teuthology.orchestra.run.vm02:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-07T10:04:41.146 DEBUG:teuthology.orchestra.run.vm10:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-07T10:04:41.149 DEBUG:teuthology.orchestra.run.vm02:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-07T10:04:41.191 DEBUG:teuthology.orchestra.run.vm02:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-07T10:04:41.235 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-07T10:04:41.235 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-07T10:04:41.284 DEBUG:teuthology.orchestra.run.vm10:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-07T10:04:41.287 DEBUG:teuthology.orchestra.run.vm10:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-07T10:04:41.333 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-07T10:04:41.333 DEBUG:teuthology.orchestra.run.vm10:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-07T10:04:41.382 DEBUG:teuthology.orchestra.run.vm02:> sudo service rsyslog restart 2026-03-07T10:04:41.383 DEBUG:teuthology.orchestra.run.vm10:> sudo service rsyslog restart 2026-03-07T10:04:41.438 INFO:teuthology.run_tasks:Running task internal.timer... 2026-03-07T10:04:41.439 INFO:teuthology.task.internal:Starting timer... 2026-03-07T10:04:41.440 INFO:teuthology.run_tasks:Running task pcp... 2026-03-07T10:04:41.443 INFO:teuthology.run_tasks:Running task selinux... 2026-03-07T10:04:41.445 INFO:teuthology.task.selinux:Excluding vm02: VMs are not yet supported 2026-03-07T10:04:41.445 INFO:teuthology.task.selinux:Excluding vm10: VMs are not yet supported 2026-03-07T10:04:41.445 DEBUG:teuthology.task.selinux:Getting current SELinux state 2026-03-07T10:04:41.445 DEBUG:teuthology.task.selinux:Existing SELinux modes: {} 2026-03-07T10:04:41.445 INFO:teuthology.task.selinux:Putting SELinux into permissive mode 2026-03-07T10:04:41.445 INFO:teuthology.run_tasks:Running task ansible.cephlab... 2026-03-07T10:04:41.446 DEBUG:teuthology.task:Applying overrides for task ansible.cephlab: {'branch': 'main', 'repo': 'https://github.com/kshtsk/ceph-cm-ansible.git', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'timezone': 'UTC'}} 2026-03-07T10:04:41.447 DEBUG:teuthology.repo_utils:Setting repo remote to https://github.com/kshtsk/ceph-cm-ansible.git 2026-03-07T10:04:41.448 INFO:teuthology.repo_utils:Fetching github.com_kshtsk_ceph-cm-ansible_main from origin 2026-03-07T10:04:42.018 DEBUG:teuthology.repo_utils:Resetting repo at /home/teuthos/src/github.com_kshtsk_ceph-cm-ansible_main to origin/main 2026-03-07T10:04:42.023 INFO:teuthology.task.ansible:Playbook: [{'import_playbook': 'ansible_managed.yml'}, {'import_playbook': 'teuthology.yml'}, {'hosts': 'testnodes', 'tasks': [{'set_fact': {'ran_from_cephlab_playbook': True}}]}, {'import_playbook': 'testnodes.yml'}, {'import_playbook': 'container-host.yml'}, {'import_playbook': 'cobbler.yml'}, {'import_playbook': 'paddles.yml'}, {'import_playbook': 'pulpito.yml'}, {'hosts': 'testnodes', 'become': True, 'tasks': [{'name': 'Touch /ceph-qa-ready', 'file': {'path': '/ceph-qa-ready', 'state': 'touch'}, 'when': 'ran_from_cephlab_playbook|bool'}]}] 2026-03-07T10:04:42.023 DEBUG:teuthology.task.ansible:Running ansible-playbook -v --extra-vars '{"ansible_ssh_user": "ubuntu", "timezone": "UTC"}' -i /tmp/teuth_ansible_inventoryezftuzrv --limit vm02.local,vm10.local /home/teuthos/src/github.com_kshtsk_ceph-cm-ansible_main/cephlab.yml --skip-tags nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs 2026-03-07T10:07:07.451 DEBUG:teuthology.task.ansible:Reconnecting to [Remote(name='ubuntu@vm02.local'), Remote(name='ubuntu@vm10.local')] 2026-03-07T10:07:07.452 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm02.local' 2026-03-07T10:07:07.452 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm02.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-07T10:07:07.513 DEBUG:teuthology.orchestra.run.vm02:> true 2026-03-07T10:07:07.740 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm02.local' 2026-03-07T10:07:07.741 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm10.local' 2026-03-07T10:07:07.741 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm10.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-07T10:07:07.800 DEBUG:teuthology.orchestra.run.vm10:> true 2026-03-07T10:07:08.029 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm10.local' 2026-03-07T10:07:08.029 INFO:teuthology.run_tasks:Running task clock... 2026-03-07T10:07:08.032 INFO:teuthology.task.clock:Syncing clocks and checking initial clock skew... 2026-03-07T10:07:08.032 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-07T10:07:08.033 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-07T10:07:08.034 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-07T10:07:08.034 DEBUG:teuthology.orchestra.run.vm10:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-07T10:07:08.048 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:08 ntpd[15624]: ntpd 4.2.8p15@1.3728-o Wed Feb 16 17:13:02 UTC 2022 (1): Starting 2026-03-07T10:07:08.048 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:08 ntpd[15624]: Command line: ntpd -gq 2026-03-07T10:07:08.048 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:08 ntpd[15624]: ---------------------------------------------------- 2026-03-07T10:07:08.048 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:08 ntpd[15624]: ntp-4 is maintained by Network Time Foundation, 2026-03-07T10:07:08.048 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:08 ntpd[15624]: Inc. (NTF), a non-profit 501(c)(3) public-benefit 2026-03-07T10:07:08.048 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:08 ntpd[15624]: corporation. Support and training for ntp-4 are 2026-03-07T10:07:08.048 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:08 ntpd[15624]: available at https://www.nwtime.org/support 2026-03-07T10:07:08.048 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:08 ntpd[15624]: ---------------------------------------------------- 2026-03-07T10:07:08.048 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:08 ntpd[15624]: proto: precision = 0.029 usec (-25) 2026-03-07T10:07:08.048 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:08 ntpd[15624]: basedate set to 2022-02-04 2026-03-07T10:07:08.048 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:08 ntpd[15624]: gps base set to 2022-02-06 (week 2196) 2026-03-07T10:07:08.048 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:08 ntpd[15624]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): good hash signature 2026-03-07T10:07:08.048 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:08 ntpd[15624]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): loaded, expire=2025-12-28T00:00:00Z last=2017-01-01T00:00:00Z ofs=37 2026-03-07T10:07:08.048 INFO:teuthology.orchestra.run.vm02.stderr: 7 Mar 10:07:08 ntpd[15624]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): expired 70 days ago 2026-03-07T10:07:08.049 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:08 ntpd[15624]: Listen and drop on 0 v6wildcard [::]:123 2026-03-07T10:07:08.049 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:08 ntpd[15624]: Listen and drop on 1 v4wildcard 0.0.0.0:123 2026-03-07T10:07:08.049 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:08 ntpd[15624]: Listen normally on 2 lo 127.0.0.1:123 2026-03-07T10:07:08.049 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:08 ntpd[15624]: Listen normally on 3 ens3 192.168.123.102:123 2026-03-07T10:07:08.049 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:08 ntpd[15624]: Listen normally on 4 lo [::1]:123 2026-03-07T10:07:08.049 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:08 ntpd[15624]: Listen normally on 5 ens3 [fe80::5055:ff:fe00:2%2]:123 2026-03-07T10:07:08.049 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:08 ntpd[15624]: Listening on routing socket on fd #22 for interface updates 2026-03-07T10:07:08.090 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:08 ntpd[15628]: ntpd 4.2.8p15@1.3728-o Wed Feb 16 17:13:02 UTC 2022 (1): Starting 2026-03-07T10:07:08.090 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:08 ntpd[15628]: Command line: ntpd -gq 2026-03-07T10:07:08.090 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:08 ntpd[15628]: ---------------------------------------------------- 2026-03-07T10:07:08.090 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:08 ntpd[15628]: ntp-4 is maintained by Network Time Foundation, 2026-03-07T10:07:08.090 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:08 ntpd[15628]: Inc. (NTF), a non-profit 501(c)(3) public-benefit 2026-03-07T10:07:08.090 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:08 ntpd[15628]: corporation. Support and training for ntp-4 are 2026-03-07T10:07:08.090 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:08 ntpd[15628]: available at https://www.nwtime.org/support 2026-03-07T10:07:08.090 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:08 ntpd[15628]: ---------------------------------------------------- 2026-03-07T10:07:08.090 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:08 ntpd[15628]: proto: precision = 0.030 usec (-25) 2026-03-07T10:07:08.090 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:08 ntpd[15628]: basedate set to 2022-02-04 2026-03-07T10:07:08.090 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:08 ntpd[15628]: gps base set to 2022-02-06 (week 2196) 2026-03-07T10:07:08.090 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:08 ntpd[15628]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): good hash signature 2026-03-07T10:07:08.090 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:08 ntpd[15628]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): loaded, expire=2025-12-28T00:00:00Z last=2017-01-01T00:00:00Z ofs=37 2026-03-07T10:07:08.091 INFO:teuthology.orchestra.run.vm10.stderr: 7 Mar 10:07:08 ntpd[15628]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): expired 70 days ago 2026-03-07T10:07:08.091 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:08 ntpd[15628]: Listen and drop on 0 v6wildcard [::]:123 2026-03-07T10:07:08.091 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:08 ntpd[15628]: Listen and drop on 1 v4wildcard 0.0.0.0:123 2026-03-07T10:07:08.091 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:08 ntpd[15628]: Listen normally on 2 lo 127.0.0.1:123 2026-03-07T10:07:08.091 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:08 ntpd[15628]: Listen normally on 3 ens3 192.168.123.110:123 2026-03-07T10:07:08.091 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:08 ntpd[15628]: Listen normally on 4 lo [::1]:123 2026-03-07T10:07:08.091 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:08 ntpd[15628]: Listen normally on 5 ens3 [fe80::5055:ff:fe00:a%2]:123 2026-03-07T10:07:08.091 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:08 ntpd[15628]: Listening on routing socket on fd #22 for interface updates 2026-03-07T10:07:09.049 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:09 ntpd[15624]: Soliciting pool server 88.198.7.62 2026-03-07T10:07:09.091 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:09 ntpd[15628]: Soliciting pool server 88.198.7.62 2026-03-07T10:07:10.048 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:10 ntpd[15624]: Soliciting pool server 130.61.89.107 2026-03-07T10:07:10.049 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:10 ntpd[15624]: Soliciting pool server 172.104.149.161 2026-03-07T10:07:10.089 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:10 ntpd[15628]: Soliciting pool server 130.61.89.107 2026-03-07T10:07:10.090 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:10 ntpd[15628]: Soliciting pool server 172.104.149.161 2026-03-07T10:07:11.048 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:11 ntpd[15624]: Soliciting pool server 5.75.181.179 2026-03-07T10:07:11.048 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:11 ntpd[15624]: Soliciting pool server 90.187.112.137 2026-03-07T10:07:11.090 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:11 ntpd[15628]: Soliciting pool server 5.75.181.179 2026-03-07T10:07:11.090 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:11 ntpd[15628]: Soliciting pool server 90.187.112.137 2026-03-07T10:07:11.165 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:11 ntpd[15624]: Soliciting pool server 188.245.170.46 2026-03-07T10:07:11.165 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:11 ntpd[15628]: Soliciting pool server 188.245.170.46 2026-03-07T10:07:12.048 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:12 ntpd[15624]: Soliciting pool server 144.76.59.37 2026-03-07T10:07:12.048 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:12 ntpd[15624]: Soliciting pool server 85.121.52.237 2026-03-07T10:07:12.049 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:12 ntpd[15624]: Soliciting pool server 85.215.229.230 2026-03-07T10:07:12.049 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:12 ntpd[15624]: Soliciting pool server 194.59.205.229 2026-03-07T10:07:12.089 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:12 ntpd[15628]: Soliciting pool server 144.76.59.37 2026-03-07T10:07:12.090 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:12 ntpd[15628]: Soliciting pool server 85.121.52.237 2026-03-07T10:07:12.090 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:12 ntpd[15628]: Soliciting pool server 85.215.229.230 2026-03-07T10:07:12.090 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:12 ntpd[15628]: Soliciting pool server 194.59.205.229 2026-03-07T10:07:13.048 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:13 ntpd[15624]: Soliciting pool server 185.216.176.59 2026-03-07T10:07:13.049 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:13 ntpd[15624]: Soliciting pool server 162.159.200.1 2026-03-07T10:07:13.049 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:13 ntpd[15624]: Soliciting pool server 134.60.111.110 2026-03-07T10:07:13.049 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:13 ntpd[15624]: Soliciting pool server 185.125.190.56 2026-03-07T10:07:13.089 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:13 ntpd[15628]: Soliciting pool server 185.216.176.59 2026-03-07T10:07:13.089 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:13 ntpd[15628]: Soliciting pool server 162.159.200.1 2026-03-07T10:07:13.090 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:13 ntpd[15628]: Soliciting pool server 134.60.111.110 2026-03-07T10:07:13.090 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:13 ntpd[15628]: Soliciting pool server 185.125.190.56 2026-03-07T10:07:14.048 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:14 ntpd[15624]: Soliciting pool server 91.189.91.157 2026-03-07T10:07:14.048 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:14 ntpd[15624]: Soliciting pool server 128.140.109.119 2026-03-07T10:07:14.049 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:14 ntpd[15624]: Soliciting pool server 195.201.20.16 2026-03-07T10:07:14.089 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:14 ntpd[15628]: Soliciting pool server 91.189.91.157 2026-03-07T10:07:14.089 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:14 ntpd[15628]: Soliciting pool server 128.140.109.119 2026-03-07T10:07:14.090 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:14 ntpd[15628]: Soliciting pool server 195.201.20.16 2026-03-07T10:07:17.112 INFO:teuthology.orchestra.run.vm10.stdout: 7 Mar 10:07:17 ntpd[15628]: ntpd: time slew -0.005913 s 2026-03-07T10:07:17.112 INFO:teuthology.orchestra.run.vm10.stdout:ntpd: time slew -0.005913s 2026-03-07T10:07:17.132 INFO:teuthology.orchestra.run.vm10.stdout: remote refid st t when poll reach delay offset jitter 2026-03-07T10:07:17.132 INFO:teuthology.orchestra.run.vm10.stdout:============================================================================== 2026-03-07T10:07:17.132 INFO:teuthology.orchestra.run.vm10.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T10:07:17.132 INFO:teuthology.orchestra.run.vm10.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T10:07:17.132 INFO:teuthology.orchestra.run.vm10.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T10:07:17.132 INFO:teuthology.orchestra.run.vm10.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T10:07:17.132 INFO:teuthology.orchestra.run.vm10.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T10:07:18.074 INFO:teuthology.orchestra.run.vm02.stdout: 7 Mar 10:07:18 ntpd[15624]: ntpd: time slew -0.000393 s 2026-03-07T10:07:18.074 INFO:teuthology.orchestra.run.vm02.stdout:ntpd: time slew -0.000393s 2026-03-07T10:07:18.096 INFO:teuthology.orchestra.run.vm02.stdout: remote refid st t when poll reach delay offset jitter 2026-03-07T10:07:18.096 INFO:teuthology.orchestra.run.vm02.stdout:============================================================================== 2026-03-07T10:07:18.096 INFO:teuthology.orchestra.run.vm02.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T10:07:18.096 INFO:teuthology.orchestra.run.vm02.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T10:07:18.096 INFO:teuthology.orchestra.run.vm02.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T10:07:18.096 INFO:teuthology.orchestra.run.vm02.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T10:07:18.096 INFO:teuthology.orchestra.run.vm02.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T10:07:18.097 INFO:teuthology.run_tasks:Running task cephadm... 2026-03-07T10:07:18.145 INFO:tasks.cephadm:Config: {'conf': {'global': {'mon election default strategy': 1}, 'mgr': {'debug mgr': 20, 'debug ms': 1, 'mgr/cephadm/use_agent': False}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', 'CEPHADM_FAILED_DAEMON'], 'log-only-match': ['CEPHADM_'], 'sha1': '340d3c24fc6ae7529322dc7ccee6c6cb2589da0a', 'cephadm_binary_url': 'https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm', 'containers': {'image': 'harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5'}} 2026-03-07T10:07:18.146 INFO:tasks.cephadm:Provided image contains tag or digest, using it as is 2026-03-07T10:07:18.146 INFO:tasks.cephadm:Cluster image is harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 2026-03-07T10:07:18.146 INFO:tasks.cephadm:Cluster fsid is 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:07:18.146 INFO:tasks.cephadm:Choosing monitor IPs and ports... 2026-03-07T10:07:18.146 INFO:tasks.cephadm:Monitor IPs: {'mon.a': '192.168.123.102', 'mon.b': '192.168.123.110'} 2026-03-07T10:07:18.146 INFO:tasks.cephadm:First mon is mon.a on vm02 2026-03-07T10:07:18.146 INFO:tasks.cephadm:First mgr is a 2026-03-07T10:07:18.146 INFO:tasks.cephadm:Normalizing hostnames... 2026-03-07T10:07:18.146 DEBUG:teuthology.orchestra.run.vm02:> sudo hostname $(hostname -s) 2026-03-07T10:07:18.155 DEBUG:teuthology.orchestra.run.vm10:> sudo hostname $(hostname -s) 2026-03-07T10:07:18.162 INFO:tasks.cephadm:Downloading cephadm from url: https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm 2026-03-07T10:07:18.163 DEBUG:teuthology.orchestra.run.vm02:> curl --silent -L https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-07T10:07:20.535 INFO:teuthology.orchestra.run.vm02.stdout:-rw-rw-r-- 1 ubuntu ubuntu 787672 Mar 7 10:07 /home/ubuntu/cephtest/cephadm 2026-03-07T10:07:20.536 DEBUG:teuthology.orchestra.run.vm10:> curl --silent -L https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-07T10:07:21.872 INFO:teuthology.orchestra.run.vm10.stdout:-rw-rw-r-- 1 ubuntu ubuntu 787672 Mar 7 10:07 /home/ubuntu/cephtest/cephadm 2026-03-07T10:07:21.872 DEBUG:teuthology.orchestra.run.vm02:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-07T10:07:21.876 DEBUG:teuthology.orchestra.run.vm10:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-07T10:07:21.882 INFO:tasks.cephadm:Pulling image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 on all hosts... 2026-03-07T10:07:21.882 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 pull 2026-03-07T10:07:21.920 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 pull 2026-03-07T10:07:22.165 INFO:teuthology.orchestra.run.vm02.stderr:Pulling container image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5... 2026-03-07T10:07:22.167 INFO:teuthology.orchestra.run.vm10.stderr:Pulling container image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5... 2026-03-07T10:07:44.270 INFO:teuthology.orchestra.run.vm02.stdout:{ 2026-03-07T10:07:44.270 INFO:teuthology.orchestra.run.vm02.stdout: "ceph_version": "ceph version 19.2.3-39-g340d3c24fc6 (340d3c24fc6ae7529322dc7ccee6c6cb2589da0a) squid (stable)", 2026-03-07T10:07:44.271 INFO:teuthology.orchestra.run.vm02.stdout: "image_id": "8bccc98d839aa18345ec1336292d0452ca331737e49f12524f635044dcabcfe1", 2026-03-07T10:07:44.271 INFO:teuthology.orchestra.run.vm02.stdout: "repo_digests": [ 2026-03-07T10:07:44.271 INFO:teuthology.orchestra.run.vm02.stdout: "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:ffa52c72fad7bdd2657408de9cf8d87fc2c72f716d1a00277ba13f7c12b404e0" 2026-03-07T10:07:44.271 INFO:teuthology.orchestra.run.vm02.stdout: ] 2026-03-07T10:07:44.271 INFO:teuthology.orchestra.run.vm02.stdout:} 2026-03-07T10:07:50.048 INFO:teuthology.orchestra.run.vm10.stdout:{ 2026-03-07T10:07:50.048 INFO:teuthology.orchestra.run.vm10.stdout: "ceph_version": "ceph version 19.2.3-39-g340d3c24fc6 (340d3c24fc6ae7529322dc7ccee6c6cb2589da0a) squid (stable)", 2026-03-07T10:07:50.048 INFO:teuthology.orchestra.run.vm10.stdout: "image_id": "8bccc98d839aa18345ec1336292d0452ca331737e49f12524f635044dcabcfe1", 2026-03-07T10:07:50.048 INFO:teuthology.orchestra.run.vm10.stdout: "repo_digests": [ 2026-03-07T10:07:50.048 INFO:teuthology.orchestra.run.vm10.stdout: "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:ffa52c72fad7bdd2657408de9cf8d87fc2c72f716d1a00277ba13f7c12b404e0" 2026-03-07T10:07:50.048 INFO:teuthology.orchestra.run.vm10.stdout: ] 2026-03-07T10:07:50.048 INFO:teuthology.orchestra.run.vm10.stdout:} 2026-03-07T10:07:50.075 DEBUG:teuthology.orchestra.run.vm02:> sudo mkdir -p /etc/ceph 2026-03-07T10:07:50.082 DEBUG:teuthology.orchestra.run.vm10:> sudo mkdir -p /etc/ceph 2026-03-07T10:07:50.090 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod 777 /etc/ceph 2026-03-07T10:07:50.131 DEBUG:teuthology.orchestra.run.vm10:> sudo chmod 777 /etc/ceph 2026-03-07T10:07:50.138 INFO:tasks.cephadm:Writing seed config... 2026-03-07T10:07:50.139 INFO:tasks.cephadm: override: [global] mon election default strategy = 1 2026-03-07T10:07:50.139 INFO:tasks.cephadm: override: [mgr] debug mgr = 20 2026-03-07T10:07:50.139 INFO:tasks.cephadm: override: [mgr] debug ms = 1 2026-03-07T10:07:50.139 INFO:tasks.cephadm: override: [mgr] mgr/cephadm/use_agent = False 2026-03-07T10:07:50.139 INFO:tasks.cephadm: override: [mon] debug mon = 20 2026-03-07T10:07:50.139 INFO:tasks.cephadm: override: [mon] debug ms = 1 2026-03-07T10:07:50.139 INFO:tasks.cephadm: override: [mon] debug paxos = 20 2026-03-07T10:07:50.139 INFO:tasks.cephadm: override: [osd] debug ms = 1 2026-03-07T10:07:50.139 INFO:tasks.cephadm: override: [osd] debug osd = 20 2026-03-07T10:07:50.139 INFO:tasks.cephadm: override: [osd] osd mclock iops capacity threshold hdd = 49000 2026-03-07T10:07:50.139 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-07T10:07:50.139 DEBUG:teuthology.orchestra.run.vm02:> dd of=/home/ubuntu/cephtest/seed.ceph.conf 2026-03-07T10:07:50.174 DEBUG:tasks.cephadm:Final config: [global] # make logging friendly to teuthology log_to_file = true log_to_stderr = false log to journald = false mon cluster log to file = true mon cluster log file level = debug mon clock drift allowed = 1.000 # replicate across OSDs, not hosts osd crush chooseleaf type = 0 #osd pool default size = 2 osd pool default erasure code profile = plugin=jerasure technique=reed_sol_van k=2 m=1 crush-failure-domain=osd # enable some debugging auth debug = true ms die on old message = true ms die on bug = true debug asserts on shutdown = true # adjust warnings mon max pg per osd = 10000# >= luminous mon pg warn max object skew = 0 mon osd allow primary affinity = true mon osd allow pg remap = true mon warn on legacy crush tunables = false mon warn on crush straw calc version zero = false mon warn on no sortbitwise = false mon warn on osd down out interval zero = false mon warn on too few osds = false mon_warn_on_pool_pg_num_not_power_of_two = false # disable pg_autoscaler by default for new pools osd_pool_default_pg_autoscale_mode = off # tests delete pools mon allow pool delete = true fsid = 6c715b7a-1a0d-11f1-b180-89615ccd948e mon election default strategy = 1 [osd] osd scrub load threshold = 5.0 osd scrub max interval = 600 osd mclock profile = high_recovery_ops osd recover clone overlap = true osd recovery max chunk = 1048576 osd deep scrub update digest min age = 30 osd map max advance = 10 osd memory target autotune = true # debugging osd debug shutdown = true osd debug op order = true osd debug verify stray on activate = true osd debug pg log writeout = true osd debug verify cached snaps = true osd debug verify missing on start = true osd debug misdirected ops = true osd op queue = debug_random osd op queue cut off = debug_random osd shutdown pgref assert = true bdev debug aio = true osd sloppy crc = true debug ms = 1 debug osd = 20 osd mclock iops capacity threshold hdd = 49000 [mgr] mon reweight min pgs per osd = 4 mon reweight min bytes per osd = 10 mgr/telemetry/nag = false debug mgr = 20 debug ms = 1 mgr/cephadm/use_agent = False [mon] mon data avail warn = 5 mon mgr mkfs grace = 240 mon reweight min pgs per osd = 4 mon osd reporter subtree level = osd mon osd prime pg temp = true mon reweight min bytes per osd = 10 # rotate auth tickets quickly to exercise renewal paths auth mon ticket ttl = 660# 11m auth service ticket ttl = 240# 4m # don't complain about global id reclaim mon_warn_on_insecure_global_id_reclaim = false mon_warn_on_insecure_global_id_reclaim_allowed = false debug mon = 20 debug ms = 1 debug paxos = 20 [client.rgw] rgw cache enabled = true rgw enable ops log = true rgw enable usage log = true 2026-03-07T10:07:50.175 DEBUG:teuthology.orchestra.run.vm02:mon.a> sudo journalctl -f -n 0 -u ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@mon.a.service 2026-03-07T10:07:50.216 DEBUG:teuthology.orchestra.run.vm02:mgr.a> sudo journalctl -f -n 0 -u ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@mgr.a.service 2026-03-07T10:07:50.260 INFO:tasks.cephadm:Bootstrapping... 2026-03-07T10:07:50.260 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 -v bootstrap --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e --config /home/ubuntu/cephtest/seed.ceph.conf --output-config /etc/ceph/ceph.conf --output-keyring /etc/ceph/ceph.client.admin.keyring --output-pub-ssh-key /home/ubuntu/cephtest/ceph.pub --mon-id a --mgr-id a --orphan-initial-daemons --skip-monitoring-stack --mon-ip 192.168.123.102 --skip-admin-label && sudo chmod +r /etc/ceph/ceph.client.admin.keyring 2026-03-07T10:07:50.555 INFO:teuthology.orchestra.run.vm02.stdout:-------------------------------------------------------------------------------- 2026-03-07T10:07:50.555 INFO:teuthology.orchestra.run.vm02.stdout:cephadm ['--image', 'harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5', '-v', 'bootstrap', '--fsid', '6c715b7a-1a0d-11f1-b180-89615ccd948e', '--config', '/home/ubuntu/cephtest/seed.ceph.conf', '--output-config', '/etc/ceph/ceph.conf', '--output-keyring', '/etc/ceph/ceph.client.admin.keyring', '--output-pub-ssh-key', '/home/ubuntu/cephtest/ceph.pub', '--mon-id', 'a', '--mgr-id', 'a', '--orphan-initial-daemons', '--skip-monitoring-stack', '--mon-ip', '192.168.123.102', '--skip-admin-label'] 2026-03-07T10:07:50.555 INFO:teuthology.orchestra.run.vm02.stderr:Specifying an fsid for your cluster offers no advantages and may increase the likelihood of fsid conflicts. 2026-03-07T10:07:50.555 INFO:teuthology.orchestra.run.vm02.stdout:Verifying podman|docker is present... 2026-03-07T10:07:50.555 INFO:teuthology.orchestra.run.vm02.stdout:Verifying lvm2 is present... 2026-03-07T10:07:50.555 INFO:teuthology.orchestra.run.vm02.stdout:Verifying time synchronization is in place... 2026-03-07T10:07:50.558 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-07T10:07:50.558 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-07T10:07:50.561 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-07T10:07:50.561 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-07T10:07:50.563 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl is-enabled chronyd.service 2026-03-07T10:07:50.563 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to get unit file state for chronyd.service: No such file or directory 2026-03-07T10:07:50.565 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active chronyd.service 2026-03-07T10:07:50.565 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-07T10:07:50.567 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl is-enabled systemd-timesyncd.service 2026-03-07T10:07:50.567 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout masked 2026-03-07T10:07:50.569 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active systemd-timesyncd.service 2026-03-07T10:07:50.569 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-07T10:07:50.572 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl is-enabled ntpd.service 2026-03-07T10:07:50.572 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to get unit file state for ntpd.service: No such file or directory 2026-03-07T10:07:50.574 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active ntpd.service 2026-03-07T10:07:50.574 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-07T10:07:50.577 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout enabled 2026-03-07T10:07:50.579 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout active 2026-03-07T10:07:50.579 INFO:teuthology.orchestra.run.vm02.stdout:Unit ntp.service is enabled and running 2026-03-07T10:07:50.579 INFO:teuthology.orchestra.run.vm02.stdout:Repeating the final host check... 2026-03-07T10:07:50.579 INFO:teuthology.orchestra.run.vm02.stdout:docker (/usr/bin/docker) is present 2026-03-07T10:07:50.579 INFO:teuthology.orchestra.run.vm02.stdout:systemctl is present 2026-03-07T10:07:50.579 INFO:teuthology.orchestra.run.vm02.stdout:lvcreate is present 2026-03-07T10:07:50.581 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-07T10:07:50.581 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-07T10:07:50.584 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-07T10:07:50.584 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-07T10:07:50.586 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl is-enabled chronyd.service 2026-03-07T10:07:50.586 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to get unit file state for chronyd.service: No such file or directory 2026-03-07T10:07:50.588 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active chronyd.service 2026-03-07T10:07:50.588 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-07T10:07:50.590 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl is-enabled systemd-timesyncd.service 2026-03-07T10:07:50.590 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout masked 2026-03-07T10:07:50.592 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active systemd-timesyncd.service 2026-03-07T10:07:50.592 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-07T10:07:50.594 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl is-enabled ntpd.service 2026-03-07T10:07:50.594 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to get unit file state for ntpd.service: No such file or directory 2026-03-07T10:07:50.597 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active ntpd.service 2026-03-07T10:07:50.597 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-07T10:07:50.599 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout enabled 2026-03-07T10:07:50.601 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout active 2026-03-07T10:07:50.602 INFO:teuthology.orchestra.run.vm02.stdout:Unit ntp.service is enabled and running 2026-03-07T10:07:50.602 INFO:teuthology.orchestra.run.vm02.stdout:Host looks OK 2026-03-07T10:07:50.602 INFO:teuthology.orchestra.run.vm02.stdout:Cluster fsid: 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:07:50.602 INFO:teuthology.orchestra.run.vm02.stdout:Acquiring lock 139674342476480 on /run/cephadm/6c715b7a-1a0d-11f1-b180-89615ccd948e.lock 2026-03-07T10:07:50.602 INFO:teuthology.orchestra.run.vm02.stdout:Lock 139674342476480 acquired on /run/cephadm/6c715b7a-1a0d-11f1-b180-89615ccd948e.lock 2026-03-07T10:07:50.602 INFO:teuthology.orchestra.run.vm02.stdout:Verifying IP 192.168.123.102 port 3300 ... 2026-03-07T10:07:50.602 INFO:teuthology.orchestra.run.vm02.stdout:Verifying IP 192.168.123.102 port 6789 ... 2026-03-07T10:07:50.602 INFO:teuthology.orchestra.run.vm02.stdout:Base mon IP(s) is [192.168.123.102:3300, 192.168.123.102:6789], mon addrv is [v2:192.168.123.102:3300,v1:192.168.123.102:6789] 2026-03-07T10:07:50.604 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout default via 192.168.123.1 dev ens3 proto dhcp src 192.168.123.102 metric 100 2026-03-07T10:07:50.604 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout 172.17.0.0/16 dev docker0 proto kernel scope link src 172.17.0.1 linkdown 2026-03-07T10:07:50.604 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout 192.168.123.0/24 dev ens3 proto kernel scope link src 192.168.123.102 metric 100 2026-03-07T10:07:50.604 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout 192.168.123.1 dev ens3 proto dhcp scope link src 192.168.123.102 metric 100 2026-03-07T10:07:50.605 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout ::1 dev lo proto kernel metric 256 pref medium 2026-03-07T10:07:50.605 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout fe80::/64 dev ens3 proto kernel metric 256 pref medium 2026-03-07T10:07:50.606 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout 1: lo: mtu 65536 state UNKNOWN qlen 1000 2026-03-07T10:07:50.606 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout inet6 ::1/128 scope host 2026-03-07T10:07:50.606 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-07T10:07:50.606 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout 2: ens3: mtu 1500 state UP qlen 1000 2026-03-07T10:07:50.606 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout inet6 fe80::5055:ff:fe00:2/64 scope link 2026-03-07T10:07:50.606 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-07T10:07:50.607 INFO:teuthology.orchestra.run.vm02.stdout:Mon IP `192.168.123.102` is in CIDR network `192.168.123.0/24` 2026-03-07T10:07:50.607 INFO:teuthology.orchestra.run.vm02.stdout:Mon IP `192.168.123.102` is in CIDR network `192.168.123.0/24` 2026-03-07T10:07:50.607 INFO:teuthology.orchestra.run.vm02.stdout:Mon IP `192.168.123.102` is in CIDR network `192.168.123.1/32` 2026-03-07T10:07:50.607 INFO:teuthology.orchestra.run.vm02.stdout:Mon IP `192.168.123.102` is in CIDR network `192.168.123.1/32` 2026-03-07T10:07:50.607 INFO:teuthology.orchestra.run.vm02.stdout:Inferred mon public CIDR from local network configuration ['192.168.123.0/24', '192.168.123.0/24', '192.168.123.1/32', '192.168.123.1/32'] 2026-03-07T10:07:50.607 INFO:teuthology.orchestra.run.vm02.stdout:Internal network (--cluster-network) has not been provided, OSD replication will default to the public_network 2026-03-07T10:07:50.607 INFO:teuthology.orchestra.run.vm02.stdout:Pulling container image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5... 2026-03-07T10:07:51.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/docker: stdout cobaltcore-storage-v19.2.3-fasttrack-5: Pulling from custom-ceph/ceph/ceph 2026-03-07T10:07:51.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/docker: stdout Digest: sha256:ffa52c72fad7bdd2657408de9cf8d87fc2c72f716d1a00277ba13f7c12b404e0 2026-03-07T10:07:51.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/docker: stdout Status: Image is up to date for harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 2026-03-07T10:07:51.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/docker: stdout harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 2026-03-07T10:07:51.362 INFO:teuthology.orchestra.run.vm02.stdout:ceph: stdout ceph version 19.2.3-39-g340d3c24fc6 (340d3c24fc6ae7529322dc7ccee6c6cb2589da0a) squid (stable) 2026-03-07T10:07:51.363 INFO:teuthology.orchestra.run.vm02.stdout:Ceph version: ceph version 19.2.3-39-g340d3c24fc6 (340d3c24fc6ae7529322dc7ccee6c6cb2589da0a) squid (stable) 2026-03-07T10:07:51.363 INFO:teuthology.orchestra.run.vm02.stdout:Extracting ceph user uid/gid from container image... 2026-03-07T10:07:51.462 INFO:teuthology.orchestra.run.vm02.stdout:stat: stdout 167 167 2026-03-07T10:07:51.462 INFO:teuthology.orchestra.run.vm02.stdout:Creating initial keys... 2026-03-07T10:07:51.636 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-authtool: stdout AQD3+KtpUJIcJBAAVBbgbKvJEWg8Un6+mY/PGQ== 2026-03-07T10:07:51.747 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-authtool: stdout AQD3+Ktp2T7oKhAANYScOa19BHNFPPaLKUbWIQ== 2026-03-07T10:07:51.861 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-authtool: stdout AQD3+KtpU3O4MRAAsgPy7aOEVHZcUsYUAXtlAQ== 2026-03-07T10:07:51.861 INFO:teuthology.orchestra.run.vm02.stdout:Creating initial monmap... 2026-03-07T10:07:51.972 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-07T10:07:51.972 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: stdout setting min_mon_release = quincy 2026-03-07T10:07:51.972 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: set fsid to 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:07:51.972 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-07T10:07:51.972 INFO:teuthology.orchestra.run.vm02.stdout:monmaptool for a [v2:192.168.123.102:3300,v1:192.168.123.102:6789] on /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-07T10:07:51.972 INFO:teuthology.orchestra.run.vm02.stdout:setting min_mon_release = quincy 2026-03-07T10:07:51.972 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: set fsid to 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:07:51.972 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-07T10:07:51.972 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-07T10:07:51.972 INFO:teuthology.orchestra.run.vm02.stdout:Creating mon... 2026-03-07T10:07:52.110 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-07T10:07:52.110 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 1 imported monmap: 2026-03-07T10:07:52.110 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr epoch 0 2026-03-07T10:07:52.110 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:07:52.110 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr last_changed 2026-03-07T10:07:51.937005+0000 2026-03-07T10:07:52.110 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr created 2026-03-07T10:07:51.937005+0000 2026-03-07T10:07:52.110 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr min_mon_release 17 (quincy) 2026-03-07T10:07:52.110 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr election_strategy: 1 2026-03-07T10:07:52.110 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-07T10:07:52.110 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T10:07:52.110 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 0 /usr/bin/ceph-mon: set fsid to 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:07:52.110 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: RocksDB version: 7.9.2 2026-03-07T10:07:52.110 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T10:07:52.110 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Git sha 0 2026-03-07T10:07:52.110 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Compile date 2026-03-06 13:52:12 2026-03-07T10:07:52.110 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: DB SUMMARY 2026-03-07T10:07:52.110 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T10:07:52.110 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: DB Session ID: 64GK9H2GX6SU7TL7WIPW 2026-03-07T10:07:52.110 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T10:07:52.110 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 0, files: 2026-03-07T10:07:52.110 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T10:07:52.110 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.error_if_exists: 0 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.create_if_missing: 1 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.paranoid_checks: 1 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.env: 0x555d2749bca0 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.info_log: 0x555d2ef7ac00 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.statistics: (nil) 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.use_fsync: 0 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.max_log_file_size: 0 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.allow_fallocate: 1 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.use_direct_reads: 0 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.db_log_dir: 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.wal_dir: 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.write_buffer_manager: 0x555d2ef715e0 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-07T10:07:52.111 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.unordered_write: 0 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.row_cache: None 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.wal_filter: None 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.two_write_queues: 0 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.wal_compression: 0 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.atomic_flush: 0 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.log_readahead_size: 0 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.max_background_jobs: 2 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.max_background_compactions: -1 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.max_subcompactions: 1 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-07T10:07:52.112 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.max_open_files: -1 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Options.max_background_flushes: -1 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Compression algorithms supported: 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: kZSTD supported: 0 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: kXpressCompression supported: 0 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: kBZip2Compression supported: 0 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: kLZ4Compression supported: 1 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: kZlibCompression supported: 1 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: kSnappyCompression supported: 1 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.049+0000 7f2a919c6d80 4 rocksdb: [db/db_impl/db_impl_open.cc:317] Creating manifest 1 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000001 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.merge_operator: 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.compaction_filter: None 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x555d2ef6d3a0) 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr cache_index_and_filter_blocks: 1 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr cache_index_and_filter_blocks_with_high_priority: 0 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-07T10:07:52.113 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr pin_top_level_index_and_filter: 1 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr index_type: 0 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr data_block_index_type: 0 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr index_shortening: 1 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr data_block_hash_table_util_ratio: 0.750000 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr checksum: 4 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr no_block_cache: 0 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr block_cache: 0x555d2ef931f0 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr block_cache_name: BinnedLRUCache 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr block_cache_options: 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr capacity : 536870912 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr num_shard_bits : 4 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr strict_capacity_limit : 0 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr high_pri_pool_ratio: 0.000 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr block_cache_compressed: (nil) 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr persistent_cache: (nil) 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr block_size: 4096 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr block_size_deviation: 10 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr block_restart_interval: 16 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr index_block_restart_interval: 1 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr metadata_block_size: 4096 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr partition_filters: 0 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr use_delta_encoding: 1 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr filter_policy: bloomfilter 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr whole_key_filtering: 1 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr verify_compression: 0 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr read_amp_bytes_per_bit: 0 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr format_version: 5 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr enable_index_compression: 1 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr block_align: 0 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr max_auto_readahead_size: 262144 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr prepopulate_block_cache: 0 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr initial_auto_readahead_size: 8192 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr num_file_reads_for_auto_readahead: 2 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.compression: NoCompression 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.num_levels: 7 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-07T10:07:52.114 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-07T10:07:52.115 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.inplace_update_support: 0 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.bloom_locality: 0 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.max_successive_merges: 0 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.ttl: 2592000 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.enable_blob_files: false 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.min_blob_size: 0 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000001 succeeded,manifest_file_number is 1, next_file_number is 3, last_sequence is 0, log_number is 0,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 0 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 0 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: ab4b5d8f-62d9-446c-92eb-b6ba735315c3 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.053+0000 7f2a919c6d80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 5 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.057+0000 7f2a919c6d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x555d2ef94e00 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.057+0000 7f2a919c6d80 4 rocksdb: DB pointer 0x555d2f078000 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.057+0000 7f2a89150640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.057+0000 7f2a89150640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr ** DB Stats ** 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Uptime(secs): 0.0 total, 0.0 interval 2026-03-07T10:07:52.116 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr ** Compaction Stats [default] ** 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr ** Compaction Stats [default] ** 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Uptime(secs): 0.0 total, 0.0 interval 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Flush(GB): cumulative 0.000, interval 0.000 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr AddFile(GB): cumulative 0.000, interval 0.000 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr AddFile(Total Files): cumulative 0, interval 0 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr AddFile(L0 Files): cumulative 0, interval 0 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr AddFile(Keys): cumulative 0, interval 0 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Block cache BinnedLRUCache@0x555d2ef931f0#7 capacity: 512.00 MB usage: 0.00 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 1.9e-05 secs_since: 0 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Block cache entry stats(count,size,portion): Misc(1,0.00 KB,0%) 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr ** File Read Latency Histogram By Level [default] ** 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.061+0000 7f2a919c6d80 4 rocksdb: [db/db_impl/db_impl.cc:496] Shutdown: canceling all background work 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.061+0000 7f2a919c6d80 4 rocksdb: [db/db_impl/db_impl.cc:704] Shutdown complete 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T10:07:52.061+0000 7f2a919c6d80 0 /usr/bin/ceph-mon: created monfs at /var/lib/ceph/mon/ceph-a for mon.a 2026-03-07T10:07:52.117 INFO:teuthology.orchestra.run.vm02.stdout:create mon.a on 2026-03-07T10:07:52.437 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /etc/systemd/system/ceph.target. 2026-03-07T10:07:52.638 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e.target → /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e.target. 2026-03-07T10:07:52.639 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph.target.wants/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e.target → /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e.target. 2026-03-07T10:07:52.830 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@mon.a 2026-03-07T10:07:52.830 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to reset failed state of unit ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@mon.a.service: Unit ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@mon.a.service not loaded. 2026-03-07T10:07:53.013 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e.target.wants/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@mon.a.service → /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service. 2026-03-07T10:07:53.023 INFO:teuthology.orchestra.run.vm02.stdout:firewalld does not appear to be present 2026-03-07T10:07:53.023 INFO:teuthology.orchestra.run.vm02.stdout:Not possible to enable service . firewalld.service is not available 2026-03-07T10:07:53.023 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mon to start... 2026-03-07T10:07:53.023 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mon... 2026-03-07T10:07:53.353 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:53 vm02 bash[16543]: cluster 2026-03-07T10:07:53.151859+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-07T10:07:53.430 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout cluster: 2026-03-07T10:07:53.430 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout id: 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:07:53.430 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout health: HEALTH_OK 2026-03-07T10:07:53.430 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-07T10:07:53.430 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout services: 2026-03-07T10:07:53.430 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon: 1 daemons, quorum a (age 0.18563s) 2026-03-07T10:07:53.430 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mgr: no daemons active 2026-03-07T10:07:53.430 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd: 0 osds: 0 up, 0 in 2026-03-07T10:07:53.431 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-07T10:07:53.431 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout data: 2026-03-07T10:07:53.431 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout pools: 0 pools, 0 pgs 2026-03-07T10:07:53.431 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout objects: 0 objects, 0 B 2026-03-07T10:07:53.431 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout usage: 0 B used, 0 B / 0 B avail 2026-03-07T10:07:53.431 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout pgs: 2026-03-07T10:07:53.431 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-07T10:07:53.431 INFO:teuthology.orchestra.run.vm02.stdout:mon is available 2026-03-07T10:07:53.431 INFO:teuthology.orchestra.run.vm02.stdout:Assimilating anything we can from ceph.conf... 2026-03-07T10:07:53.765 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-07T10:07:53.766 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [global] 2026-03-07T10:07:53.766 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout fsid = 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:07:53.766 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-07T10:07:53.766 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.102:3300,v1:192.168.123.102:6789] 2026-03-07T10:07:53.766 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-07T10:07:53.766 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-07T10:07:53.766 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-07T10:07:53.766 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-07T10:07:53.766 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-07T10:07:53.766 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-07T10:07:53.766 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mgr/cephadm/use_agent = False 2026-03-07T10:07:53.766 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-07T10:07:53.766 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-07T10:07:53.766 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [osd] 2026-03-07T10:07:53.766 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-07T10:07:53.766 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-07T10:07:53.766 INFO:teuthology.orchestra.run.vm02.stdout:Generating new minimal ceph.conf... 2026-03-07T10:07:54.085 INFO:teuthology.orchestra.run.vm02.stdout:Restarting the monitor... 2026-03-07T10:07:54.264 INFO:teuthology.orchestra.run.vm02.stdout:Setting public_network to 192.168.123.1/32,192.168.123.0/24 in mon config section 2026-03-07T10:07:54.328 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 systemd[1]: Stopping Ceph mon.a for 6c715b7a-1a0d-11f1-b180-89615ccd948e... 2026-03-07T10:07:54.328 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[16543]: debug 2026-03-07T10:07:54.121+0000 7f2ea8d82640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-07T10:07:54.328 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[16543]: debug 2026-03-07T10:07:54.121+0000 7f2ea8d82640 -1 mon.a@0(leader) e1 *** Got Signal Terminated *** 2026-03-07T10:07:54.328 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[16927]: ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e-mon-a 2026-03-07T10:07:54.328 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 systemd[1]: ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@mon.a.service: Deactivated successfully. 2026-03-07T10:07:54.328 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 systemd[1]: Stopped Ceph mon.a for 6c715b7a-1a0d-11f1-b180-89615ccd948e. 2026-03-07T10:07:54.328 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 systemd[1]: Started Ceph mon.a for 6c715b7a-1a0d-11f1-b180-89615ccd948e. 2026-03-07T10:07:54.602 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.389+0000 7f5cfa95bd80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-07T10:07:54.602 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.389+0000 7f5cfa95bd80 0 ceph version 19.2.3-39-g340d3c24fc6 (340d3c24fc6ae7529322dc7ccee6c6cb2589da0a) squid (stable), process ceph-mon, pid 7 2026-03-07T10:07:54.602 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.389+0000 7f5cfa95bd80 0 pidfile_write: ignore empty --pid-file 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 0 load: jerasure load: lrc 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: RocksDB version: 7.9.2 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Git sha 0 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Compile date 2026-03-06 13:52:12 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: DB SUMMARY 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: DB Session ID: 9F05LK2RPC57Q3PJP0B5 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: CURRENT file: CURRENT 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: MANIFEST file: MANIFEST-000010 size: 179 Bytes 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 1, files: 000008.sst 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 000009.log size: 75651 ; 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.error_if_exists: 0 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.create_if_missing: 0 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.paranoid_checks: 1 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.env: 0x55a2de93bca0 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.info_log: 0x55a2f4fea280 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.statistics: (nil) 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.use_fsync: 0 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.max_log_file_size: 0 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.allow_fallocate: 1 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.use_direct_reads: 0 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.db_log_dir: 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.wal_dir: 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.write_buffer_manager: 0x55a2f4fef900 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.unordered_write: 0 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.row_cache: None 2026-03-07T10:07:54.603 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.wal_filter: None 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.two_write_queues: 0 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.wal_compression: 0 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.atomic_flush: 0 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.log_readahead_size: 0 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.max_background_jobs: 2 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.max_background_compactions: -1 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.max_subcompactions: 1 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.max_open_files: -1 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.max_background_flushes: -1 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Compression algorithms supported: 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: kZSTD supported: 0 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: kXpressCompression supported: 0 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: kBZip2Compression supported: 0 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: kLZ4Compression supported: 1 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: kZlibCompression supported: 1 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: kSnappyCompression supported: 1 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.merge_operator: 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.compaction_filter: None 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55a2f4fea3a0) 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: cache_index_and_filter_blocks: 1 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: pin_top_level_index_and_filter: 1 2026-03-07T10:07:54.604 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: index_type: 0 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: data_block_index_type: 0 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: index_shortening: 1 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: data_block_hash_table_util_ratio: 0.750000 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: checksum: 4 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: no_block_cache: 0 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: block_cache: 0x55a2f50111f0 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: block_cache_name: BinnedLRUCache 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: block_cache_options: 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: capacity : 536870912 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: num_shard_bits : 4 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: strict_capacity_limit : 0 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: high_pri_pool_ratio: 0.000 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: block_cache_compressed: (nil) 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: persistent_cache: (nil) 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: block_size: 4096 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: block_size_deviation: 10 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: block_restart_interval: 16 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: index_block_restart_interval: 1 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: metadata_block_size: 4096 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: partition_filters: 0 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: use_delta_encoding: 1 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: filter_policy: bloomfilter 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: whole_key_filtering: 1 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: verify_compression: 0 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: read_amp_bytes_per_bit: 0 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: format_version: 5 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: enable_index_compression: 1 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: block_align: 0 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: max_auto_readahead_size: 262144 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: prepopulate_block_cache: 0 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: initial_auto_readahead_size: 8192 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: num_file_reads_for_auto_readahead: 2 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.compression: NoCompression 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.num_levels: 7 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-07T10:07:54.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.inplace_update_support: 0 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.bloom_locality: 0 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.max_successive_merges: 0 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.ttl: 2592000 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.enable_blob_files: false 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.min_blob_size: 0 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.393+0000 7f5cfa95bd80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.401+0000 7f5cfa95bd80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 succeeded,manifest_file_number is 10, next_file_number is 12, last_sequence is 5, log_number is 5,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 5 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.401+0000 7f5cfa95bd80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 5 2026-03-07T10:07:54.606 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.401+0000 7f5cfa95bd80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: ab4b5d8f-62d9-446c-92eb-b6ba735315c3 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.401+0000 7f5cfa95bd80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1772878074404500, "job": 1, "event": "recovery_started", "wal_files": [9]} 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.401+0000 7f5cfa95bd80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #9 mode 2 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.401+0000 7f5cfa95bd80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1772878074406525, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 13, "file_size": 72732, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 8, "largest_seqno": 225, "table_properties": {"data_size": 71011, "index_size": 174, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 517, "raw_key_size": 9705, "raw_average_key_size": 49, "raw_value_size": 65490, "raw_average_value_size": 334, "num_data_blocks": 8, "num_entries": 196, "num_filter_entries": 196, "num_deletions": 3, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1772878074, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "ab4b5d8f-62d9-446c-92eb-b6ba735315c3", "db_session_id": "9F05LK2RPC57Q3PJP0B5", "orig_file_number": 13, "seqno_to_time_mapping": "N/A"}} 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.401+0000 7f5cfa95bd80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1772878074406606, "job": 1, "event": "recovery_finished"} 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.401+0000 7f5cfa95bd80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 15 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.409+0000 7f5cfa95bd80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-a/store.db/000009.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.409+0000 7f5cfa95bd80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55a2f5012e00 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.409+0000 7f5cfa95bd80 4 rocksdb: DB pointer 0x55a2f512a000 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.409+0000 7f5cf0725640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.409+0000 7f5cf0725640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: ** DB Stats ** 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: ** Compaction Stats [default] ** 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: L0 2/0 72.88 KB 0.5 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 38.2 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: Sum 2/0 72.88 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 38.2 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 38.2 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: ** Compaction Stats [default] ** 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 38.2 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: Flush(GB): cumulative 0.000, interval 0.000 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: AddFile(Total Files): cumulative 0, interval 0 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: AddFile(L0 Files): cumulative 0, interval 0 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: AddFile(Keys): cumulative 0, interval 0 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: Cumulative compaction: 0.00 GB write, 4.64 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: Interval compaction: 0.00 GB write, 4.64 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: Block cache BinnedLRUCache@0x55a2f50111f0#7 capacity: 512.00 MB usage: 26.17 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 8e-06 secs_since: 0 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: Block cache entry stats(count,size,portion): DataBlock(3,25.11 KB,0.00478923%) FilterBlock(2,0.70 KB,0.00013411%) IndexBlock(2,0.36 KB,6.85453e-05%) Misc(1,0.00 KB,0%) 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: ** File Read Latency Histogram By Level [default] ** 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.413+0000 7f5cfa95bd80 0 starting mon.a rank 0 at public addrs [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] at bind addrs [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon_data /var/lib/ceph/mon/ceph-a fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.413+0000 7f5cfa95bd80 1 mon.a@-1(???) e1 preinit fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.413+0000 7f5cfa95bd80 0 mon.a@-1(???).mds e1 new map 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.413+0000 7f5cfa95bd80 0 mon.a@-1(???).mds e1 print_map 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: e1 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: btime 2026-03-07T10:07:53:157919+0000 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: enable_multiple, ever_enabled_multiple: 1,1 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: legacy client fscid: -1 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: 2026-03-07T10:07:54.607 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: No filesystems configured 2026-03-07T10:07:54.608 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.413+0000 7f5cfa95bd80 0 mon.a@-1(???).osd e1 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-07T10:07:54.608 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.413+0000 7f5cfa95bd80 0 mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-07T10:07:54.608 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.413+0000 7f5cfa95bd80 0 mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-07T10:07:54.608 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.413+0000 7f5cfa95bd80 0 mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-07T10:07:54.608 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: debug 2026-03-07T10:07:54.413+0000 7f5cfa95bd80 1 mon.a@-1(???).paxosservice(auth 1..2) refresh upgraded, format 0 -> 3 2026-03-07T10:07:54.608 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: cluster 2026-03-07T10:07:54.423777+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-07T10:07:54.608 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: cluster 2026-03-07T10:07:54.423777+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-07T10:07:54.608 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: cluster 2026-03-07T10:07:54.423846+0000 mon.a (mon.0) 2 : cluster [DBG] monmap epoch 1 2026-03-07T10:07:54.608 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: cluster 2026-03-07T10:07:54.423846+0000 mon.a (mon.0) 2 : cluster [DBG] monmap epoch 1 2026-03-07T10:07:54.608 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: cluster 2026-03-07T10:07:54.423854+0000 mon.a (mon.0) 3 : cluster [DBG] fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:07:54.608 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: cluster 2026-03-07T10:07:54.423854+0000 mon.a (mon.0) 3 : cluster [DBG] fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:07:54.608 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: cluster 2026-03-07T10:07:54.423858+0000 mon.a (mon.0) 4 : cluster [DBG] last_changed 2026-03-07T10:07:51.937005+0000 2026-03-07T10:07:54.608 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: cluster 2026-03-07T10:07:54.423858+0000 mon.a (mon.0) 4 : cluster [DBG] last_changed 2026-03-07T10:07:51.937005+0000 2026-03-07T10:07:54.608 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: cluster 2026-03-07T10:07:54.423869+0000 mon.a (mon.0) 5 : cluster [DBG] created 2026-03-07T10:07:51.937005+0000 2026-03-07T10:07:54.608 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: cluster 2026-03-07T10:07:54.423869+0000 mon.a (mon.0) 5 : cluster [DBG] created 2026-03-07T10:07:51.937005+0000 2026-03-07T10:07:54.608 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: cluster 2026-03-07T10:07:54.423875+0000 mon.a (mon.0) 6 : cluster [DBG] min_mon_release 19 (squid) 2026-03-07T10:07:54.608 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: cluster 2026-03-07T10:07:54.423875+0000 mon.a (mon.0) 6 : cluster [DBG] min_mon_release 19 (squid) 2026-03-07T10:07:54.608 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: cluster 2026-03-07T10:07:54.423879+0000 mon.a (mon.0) 7 : cluster [DBG] election_strategy: 1 2026-03-07T10:07:54.608 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: cluster 2026-03-07T10:07:54.423879+0000 mon.a (mon.0) 7 : cluster [DBG] election_strategy: 1 2026-03-07T10:07:54.608 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: cluster 2026-03-07T10:07:54.423883+0000 mon.a (mon.0) 8 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-07T10:07:54.608 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: cluster 2026-03-07T10:07:54.423883+0000 mon.a (mon.0) 8 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-07T10:07:54.608 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: cluster 2026-03-07T10:07:54.424237+0000 mon.a (mon.0) 9 : cluster [DBG] fsmap 2026-03-07T10:07:54.608 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: cluster 2026-03-07T10:07:54.424237+0000 mon.a (mon.0) 9 : cluster [DBG] fsmap 2026-03-07T10:07:54.608 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: cluster 2026-03-07T10:07:54.424256+0000 mon.a (mon.0) 10 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-07T10:07:54.608 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: cluster 2026-03-07T10:07:54.424256+0000 mon.a (mon.0) 10 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-07T10:07:54.608 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: cluster 2026-03-07T10:07:54.424938+0000 mon.a (mon.0) 11 : cluster [DBG] mgrmap e1: no daemons active 2026-03-07T10:07:54.608 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 bash[17011]: cluster 2026-03-07T10:07:54.424938+0000 mon.a (mon.0) 11 : cluster [DBG] mgrmap e1: no daemons active 2026-03-07T10:07:54.637 INFO:teuthology.orchestra.run.vm02.stdout:Wrote config to /etc/ceph/ceph.conf 2026-03-07T10:07:54.639 INFO:teuthology.orchestra.run.vm02.stdout:Wrote keyring to /etc/ceph/ceph.client.admin.keyring 2026-03-07T10:07:54.639 INFO:teuthology.orchestra.run.vm02.stdout:Creating mgr... 2026-03-07T10:07:54.639 INFO:teuthology.orchestra.run.vm02.stdout:Verifying port 0.0.0.0:9283 ... 2026-03-07T10:07:54.639 INFO:teuthology.orchestra.run.vm02.stdout:Verifying port 0.0.0.0:8765 ... 2026-03-07T10:07:54.821 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@mgr.a 2026-03-07T10:07:54.821 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to reset failed state of unit ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@mgr.a.service: Unit ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@mgr.a.service not loaded. 2026-03-07T10:07:54.930 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:07:54.985 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e.target.wants/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@mgr.a.service → /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service. 2026-03-07T10:07:54.992 INFO:teuthology.orchestra.run.vm02.stdout:firewalld does not appear to be present 2026-03-07T10:07:54.992 INFO:teuthology.orchestra.run.vm02.stdout:Not possible to enable service . firewalld.service is not available 2026-03-07T10:07:54.992 INFO:teuthology.orchestra.run.vm02.stdout:firewalld does not appear to be present 2026-03-07T10:07:54.992 INFO:teuthology.orchestra.run.vm02.stdout:Not possible to open ports <[9283, 8765]>. firewalld.service is not available 2026-03-07T10:07:54.992 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mgr to start... 2026-03-07T10:07:54.992 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mgr... 2026-03-07T10:07:55.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:54 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsid": "6c715b7a-1a0d-11f1-b180-89615ccd948e", 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "health": { 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 0 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "a" 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_age": 0, 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "btime": "2026-03-07T10:07:53:157919+0000", 2026-03-07T10:07:55.376 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-07T10:07:55.377 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-07T10:07:55.377 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:07:55.377 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-07T10:07:55.377 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-07T10:07:55.377 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-07T10:07:55.377 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-07T10:07:55.377 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-07T10:07:55.377 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-07T10:07:55.377 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "restful" 2026-03-07T10:07:55.377 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-07T10:07:55.377 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-07T10:07:55.377 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:07:55.377 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-07T10:07:55.377 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T10:07:55.377 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modified": "2026-03-07T10:07:53.158907+0000", 2026-03-07T10:07:55.377 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-07T10:07:55.377 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:07:55.377 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-07T10:07:55.377 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-07T10:07:55.377 INFO:teuthology.orchestra.run.vm02.stdout:mgr not available, waiting (1/15)... 2026-03-07T10:07:55.592 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:07:55 vm02 bash[17284]: debug 2026-03-07T10:07:55.481+0000 7ff3f19a1100 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-07T10:07:55.870 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:55 vm02 bash[17011]: audit 2026-03-07T10:07:54.589877+0000 mon.a (mon.0) 12 : audit [INF] from='client.? 192.168.123.102:0/1543476094' entity='client.admin' 2026-03-07T10:07:55.870 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:55 vm02 bash[17011]: audit 2026-03-07T10:07:54.589877+0000 mon.a (mon.0) 12 : audit [INF] from='client.? 192.168.123.102:0/1543476094' entity='client.admin' 2026-03-07T10:07:55.870 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:55 vm02 bash[17011]: audit 2026-03-07T10:07:55.308239+0000 mon.a (mon.0) 13 : audit [DBG] from='client.? 192.168.123.102:0/3121033626' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-07T10:07:55.870 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:55 vm02 bash[17011]: audit 2026-03-07T10:07:55.308239+0000 mon.a (mon.0) 13 : audit [DBG] from='client.? 192.168.123.102:0/3121033626' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-07T10:07:56.212 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:07:55 vm02 bash[17284]: debug 2026-03-07T10:07:55.865+0000 7ff3f19a1100 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-07T10:07:57.397 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:07:56 vm02 bash[17284]: debug 2026-03-07T10:07:56.977+0000 7ff3f19a1100 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-07T10:07:57.752 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-07T10:07:57.752 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-07T10:07:57.752 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsid": "6c715b7a-1a0d-11f1-b180-89615ccd948e", 2026-03-07T10:07:57.752 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "health": { 2026-03-07T10:07:57.752 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-07T10:07:57.752 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-07T10:07:57.752 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-07T10:07:57.752 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:07:57.753 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-07T10:07:57.753 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-07T10:07:57.753 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 0 2026-03-07T10:07:57.753 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-07T10:07:57.753 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-07T10:07:57.753 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "a" 2026-03-07T10:07:57.753 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-07T10:07:57.753 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_age": 3, 2026-03-07T10:07:57.753 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-07T10:07:57.753 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T10:07:57.753 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-07T10:07:57.753 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-07T10:07:57.753 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:07:57.753 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-07T10:07:57.753 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T10:07:57.753 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-07T10:07:57.753 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-07T10:07:57.753 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-07T10:07:57.753 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-07T10:07:57.753 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-07T10:07:57.753 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-07T10:07:57.754 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:07:57.754 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-07T10:07:57.754 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-07T10:07:57.754 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-07T10:07:57.754 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-07T10:07:57.754 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-07T10:07:57.754 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-07T10:07:57.754 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-07T10:07:57.754 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-07T10:07:57.754 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-07T10:07:57.754 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:07:57.754 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-07T10:07:57.754 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T10:07:57.754 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "btime": "2026-03-07T10:07:53:157919+0000", 2026-03-07T10:07:57.754 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-07T10:07:57.754 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-07T10:07:57.754 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:07:57.754 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-07T10:07:57.754 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-07T10:07:57.754 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-07T10:07:57.754 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-07T10:07:57.754 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-07T10:07:57.754 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-07T10:07:57.755 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "restful" 2026-03-07T10:07:57.755 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-07T10:07:57.755 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-07T10:07:57.755 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:07:57.755 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-07T10:07:57.755 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T10:07:57.755 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modified": "2026-03-07T10:07:53.158907+0000", 2026-03-07T10:07:57.755 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-07T10:07:57.755 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:07:57.755 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-07T10:07:57.755 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-07T10:07:57.755 INFO:teuthology.orchestra.run.vm02.stdout:mgr not available, waiting (2/15)... 2026-03-07T10:07:57.961 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:57 vm02 bash[17011]: audit 2026-03-07T10:07:57.680464+0000 mon.a (mon.0) 14 : audit [DBG] from='client.? 192.168.123.102:0/4027781281' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-07T10:07:57.961 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:07:57 vm02 bash[17011]: audit 2026-03-07T10:07:57.680464+0000 mon.a (mon.0) 14 : audit [DBG] from='client.? 192.168.123.102:0/4027781281' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-07T10:07:58.670 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:07:58 vm02 bash[17284]: debug 2026-03-07T10:07:58.329+0000 7ff3f19a1100 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-07T10:07:58.961 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:07:58 vm02 bash[17284]: debug 2026-03-07T10:07:58.665+0000 7ff3f19a1100 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-07T10:07:59.485 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:07:59 vm02 bash[17284]: debug 2026-03-07T10:07:59.217+0000 7ff3f19a1100 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-07T10:07:59.485 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:07:59 vm02 bash[17284]: debug 2026-03-07T10:07:59.341+0000 7ff3f19a1100 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-07T10:07:59.771 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:07:59 vm02 bash[17284]: debug 2026-03-07T10:07:59.481+0000 7ff3f19a1100 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-07T10:07:59.771 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:07:59 vm02 bash[17284]: debug 2026-03-07T10:07:59.641+0000 7ff3f19a1100 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-07T10:08:00.088 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:07:59 vm02 bash[17284]: debug 2026-03-07T10:07:59.829+0000 7ff3f19a1100 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsid": "6c715b7a-1a0d-11f1-b180-89615ccd948e", 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "health": { 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 0 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "a" 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_age": 5, 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-07T10:08:00.127 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-07T10:08:00.128 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:08:00.128 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-07T10:08:00.128 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-07T10:08:00.128 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-07T10:08:00.128 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-07T10:08:00.128 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-07T10:08:00.128 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-07T10:08:00.128 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-07T10:08:00.128 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-07T10:08:00.128 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-07T10:08:00.128 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:08:00.128 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-07T10:08:00.128 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T10:08:00.128 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "btime": "2026-03-07T10:07:53:157919+0000", 2026-03-07T10:08:00.128 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-07T10:08:00.129 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-07T10:08:00.129 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:08:00.129 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-07T10:08:00.129 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-07T10:08:00.129 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-07T10:08:00.129 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-07T10:08:00.129 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-07T10:08:00.129 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-07T10:08:00.129 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "restful" 2026-03-07T10:08:00.129 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-07T10:08:00.129 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-07T10:08:00.129 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:08:00.129 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-07T10:08:00.129 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T10:08:00.129 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modified": "2026-03-07T10:07:53.158907+0000", 2026-03-07T10:08:00.129 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-07T10:08:00.129 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:08:00.129 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-07T10:08:00.129 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-07T10:08:00.129 INFO:teuthology.orchestra.run.vm02.stdout:mgr not available, waiting (3/15)... 2026-03-07T10:08:00.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:00 vm02 bash[17011]: audit 2026-03-07T10:08:00.055185+0000 mon.a (mon.0) 15 : audit [DBG] from='client.? 192.168.123.102:0/1734778232' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-07T10:08:00.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:00 vm02 bash[17011]: audit 2026-03-07T10:08:00.055185+0000 mon.a (mon.0) 15 : audit [DBG] from='client.? 192.168.123.102:0/1734778232' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-07T10:08:01.400 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:01 vm02 bash[17284]: debug 2026-03-07T10:08:01.133+0000 7ff3f19a1100 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-07T10:08:01.400 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:01 vm02 bash[17284]: debug 2026-03-07T10:08:01.265+0000 7ff3f19a1100 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-07T10:08:01.711 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:01 vm02 bash[17284]: debug 2026-03-07T10:08:01.393+0000 7ff3f19a1100 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-07T10:08:02.191 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:01 vm02 bash[17284]: debug 2026-03-07T10:08:01.917+0000 7ff3f19a1100 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-07T10:08:02.191 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:02 vm02 bash[17284]: debug 2026-03-07T10:08:02.081+0000 7ff3f19a1100 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-07T10:08:02.443 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:02 vm02 bash[17284]: debug 2026-03-07T10:08:02.265+0000 7ff3f19a1100 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-07T10:08:02.484 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-07T10:08:02.484 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-07T10:08:02.484 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsid": "6c715b7a-1a0d-11f1-b180-89615ccd948e", 2026-03-07T10:08:02.484 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "health": { 2026-03-07T10:08:02.484 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-07T10:08:02.484 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-07T10:08:02.484 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-07T10:08:02.485 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:08:02.485 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-07T10:08:02.485 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-07T10:08:02.485 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 0 2026-03-07T10:08:02.485 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-07T10:08:02.485 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-07T10:08:02.485 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "a" 2026-03-07T10:08:02.485 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-07T10:08:02.485 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_age": 7, 2026-03-07T10:08:02.485 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-07T10:08:02.485 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T10:08:02.485 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-07T10:08:02.485 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-07T10:08:02.485 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:08:02.485 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-07T10:08:02.485 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T10:08:02.485 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-07T10:08:02.485 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-07T10:08:02.485 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-07T10:08:02.485 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-07T10:08:02.485 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-07T10:08:02.485 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "btime": "2026-03-07T10:07:53:157919+0000", 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "restful" 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modified": "2026-03-07T10:07:53.158907+0000", 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-07T10:08:02.486 INFO:teuthology.orchestra.run.vm02.stdout:mgr not available, waiting (4/15)... 2026-03-07T10:08:02.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:02 vm02 bash[17011]: audit 2026-03-07T10:08:02.422086+0000 mon.a (mon.0) 16 : audit [DBG] from='client.? 192.168.123.102:0/3620922213' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-07T10:08:02.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:02 vm02 bash[17011]: audit 2026-03-07T10:08:02.422086+0000 mon.a (mon.0) 16 : audit [DBG] from='client.? 192.168.123.102:0/3620922213' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-07T10:08:02.712 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:02 vm02 bash[17284]: debug 2026-03-07T10:08:02.645+0000 7ff3f19a1100 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-07T10:08:03.461 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:03 vm02 bash[17284]: debug 2026-03-07T10:08:03.137+0000 7ff3f19a1100 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-07T10:08:03.954 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:03 vm02 bash[17284]: debug 2026-03-07T10:08:03.681+0000 7ff3f19a1100 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-07T10:08:03.955 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:03 vm02 bash[17284]: debug 2026-03-07T10:08:03.813+0000 7ff3f19a1100 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-07T10:08:04.211 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:03 vm02 bash[17284]: debug 2026-03-07T10:08:03.949+0000 7ff3f19a1100 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-07T10:08:04.711 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:04 vm02 bash[17284]: debug 2026-03-07T10:08:04.437+0000 7ff3f19a1100 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-07T10:08:04.848 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsid": "6c715b7a-1a0d-11f1-b180-89615ccd948e", 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "health": { 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 0 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "a" 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_age": 10, 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-07T10:08:04.849 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T10:08:04.850 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "btime": "2026-03-07T10:07:53:157919+0000", 2026-03-07T10:08:04.850 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-07T10:08:04.850 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-07T10:08:04.850 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:08:04.850 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-07T10:08:04.850 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-07T10:08:04.850 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-07T10:08:04.850 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-07T10:08:04.850 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-07T10:08:04.850 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-07T10:08:04.850 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "restful" 2026-03-07T10:08:04.850 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-07T10:08:04.850 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-07T10:08:04.850 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:08:04.850 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-07T10:08:04.850 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T10:08:04.850 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modified": "2026-03-07T10:07:53.158907+0000", 2026-03-07T10:08:04.850 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-07T10:08:04.850 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:08:04.850 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-07T10:08:04.850 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-07T10:08:04.850 INFO:teuthology.orchestra.run.vm02.stdout:mgr not available, waiting (5/15)... 2026-03-07T10:08:05.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:04 vm02 bash[17011]: audit 2026-03-07T10:08:04.786456+0000 mon.a (mon.0) 17 : audit [DBG] from='client.? 192.168.123.102:0/1215826062' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-07T10:08:05.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:04 vm02 bash[17011]: audit 2026-03-07T10:08:04.786456+0000 mon.a (mon.0) 17 : audit [DBG] from='client.? 192.168.123.102:0/1215826062' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-07T10:08:05.881 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:05 vm02 bash[17284]: debug 2026-03-07T10:08:05.529+0000 7ff3f19a1100 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-07T10:08:06.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:05 vm02 bash[17011]: cluster 2026-03-07T10:08:05.537590+0000 mon.a (mon.0) 18 : cluster [INF] Activating manager daemon a 2026-03-07T10:08:06.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:05 vm02 bash[17011]: cluster 2026-03-07T10:08:05.537590+0000 mon.a (mon.0) 18 : cluster [INF] Activating manager daemon a 2026-03-07T10:08:06.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:05 vm02 bash[17011]: cluster 2026-03-07T10:08:05.572534+0000 mon.a (mon.0) 19 : cluster [DBG] mgrmap e2: a(active, starting, since 0.0350842s) 2026-03-07T10:08:06.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:05 vm02 bash[17011]: cluster 2026-03-07T10:08:05.572534+0000 mon.a (mon.0) 19 : cluster [DBG] mgrmap e2: a(active, starting, since 0.0350842s) 2026-03-07T10:08:06.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:05 vm02 bash[17011]: audit 2026-03-07T10:08:05.574614+0000 mon.a (mon.0) 20 : audit [DBG] from='mgr.14100 192.168.123.102:0/138859343' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-07T10:08:06.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:05 vm02 bash[17011]: audit 2026-03-07T10:08:05.574614+0000 mon.a (mon.0) 20 : audit [DBG] from='mgr.14100 192.168.123.102:0/138859343' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-07T10:08:06.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:05 vm02 bash[17011]: audit 2026-03-07T10:08:05.574953+0000 mon.a (mon.0) 21 : audit [DBG] from='mgr.14100 192.168.123.102:0/138859343' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-07T10:08:06.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:05 vm02 bash[17011]: audit 2026-03-07T10:08:05.574953+0000 mon.a (mon.0) 21 : audit [DBG] from='mgr.14100 192.168.123.102:0/138859343' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-07T10:08:06.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:05 vm02 bash[17011]: audit 2026-03-07T10:08:05.575281+0000 mon.a (mon.0) 22 : audit [DBG] from='mgr.14100 192.168.123.102:0/138859343' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-07T10:08:06.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:05 vm02 bash[17011]: audit 2026-03-07T10:08:05.575281+0000 mon.a (mon.0) 22 : audit [DBG] from='mgr.14100 192.168.123.102:0/138859343' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-07T10:08:06.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:05 vm02 bash[17011]: audit 2026-03-07T10:08:05.575698+0000 mon.a (mon.0) 23 : audit [DBG] from='mgr.14100 192.168.123.102:0/138859343' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-07T10:08:06.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:05 vm02 bash[17011]: audit 2026-03-07T10:08:05.575698+0000 mon.a (mon.0) 23 : audit [DBG] from='mgr.14100 192.168.123.102:0/138859343' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-07T10:08:06.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:05 vm02 bash[17011]: audit 2026-03-07T10:08:05.576000+0000 mon.a (mon.0) 24 : audit [DBG] from='mgr.14100 192.168.123.102:0/138859343' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-07T10:08:06.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:05 vm02 bash[17011]: audit 2026-03-07T10:08:05.576000+0000 mon.a (mon.0) 24 : audit [DBG] from='mgr.14100 192.168.123.102:0/138859343' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-07T10:08:06.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:05 vm02 bash[17011]: cluster 2026-03-07T10:08:05.582803+0000 mon.a (mon.0) 25 : cluster [INF] Manager daemon a is now available 2026-03-07T10:08:06.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:05 vm02 bash[17011]: cluster 2026-03-07T10:08:05.582803+0000 mon.a (mon.0) 25 : cluster [INF] Manager daemon a is now available 2026-03-07T10:08:06.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:05 vm02 bash[17011]: audit 2026-03-07T10:08:05.631726+0000 mon.a (mon.0) 26 : audit [INF] from='mgr.14100 192.168.123.102:0/138859343' entity='mgr.a' 2026-03-07T10:08:06.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:05 vm02 bash[17011]: audit 2026-03-07T10:08:05.631726+0000 mon.a (mon.0) 26 : audit [INF] from='mgr.14100 192.168.123.102:0/138859343' entity='mgr.a' 2026-03-07T10:08:06.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:05 vm02 bash[17011]: audit 2026-03-07T10:08:05.633296+0000 mon.a (mon.0) 27 : audit [INF] from='mgr.14100 192.168.123.102:0/138859343' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-07T10:08:06.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:05 vm02 bash[17011]: audit 2026-03-07T10:08:05.633296+0000 mon.a (mon.0) 27 : audit [INF] from='mgr.14100 192.168.123.102:0/138859343' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-07T10:08:06.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:05 vm02 bash[17011]: audit 2026-03-07T10:08:05.634753+0000 mon.a (mon.0) 28 : audit [INF] from='mgr.14100 192.168.123.102:0/138859343' entity='mgr.a' 2026-03-07T10:08:06.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:05 vm02 bash[17011]: audit 2026-03-07T10:08:05.634753+0000 mon.a (mon.0) 28 : audit [INF] from='mgr.14100 192.168.123.102:0/138859343' entity='mgr.a' 2026-03-07T10:08:06.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:05 vm02 bash[17011]: audit 2026-03-07T10:08:05.638725+0000 mon.a (mon.0) 29 : audit [INF] from='mgr.14100 192.168.123.102:0/138859343' entity='mgr.a' 2026-03-07T10:08:06.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:05 vm02 bash[17011]: audit 2026-03-07T10:08:05.638725+0000 mon.a (mon.0) 29 : audit [INF] from='mgr.14100 192.168.123.102:0/138859343' entity='mgr.a' 2026-03-07T10:08:06.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:05 vm02 bash[17011]: audit 2026-03-07T10:08:05.639260+0000 mon.a (mon.0) 30 : audit [INF] from='mgr.14100 192.168.123.102:0/138859343' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-07T10:08:06.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:05 vm02 bash[17011]: audit 2026-03-07T10:08:05.639260+0000 mon.a (mon.0) 30 : audit [INF] from='mgr.14100 192.168.123.102:0/138859343' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsid": "6c715b7a-1a0d-11f1-b180-89615ccd948e", 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "health": { 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 0 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "a" 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_age": 12, 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-07T10:08:07.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-07T10:08:07.291 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:08:07.291 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-07T10:08:07.291 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-07T10:08:07.291 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-07T10:08:07.291 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-07T10:08:07.291 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-07T10:08:07.291 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-07T10:08:07.291 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-07T10:08:07.291 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-07T10:08:07.291 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-07T10:08:07.291 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:08:07.291 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-07T10:08:07.291 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T10:08:07.291 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "btime": "2026-03-07T10:07:53:157919+0000", 2026-03-07T10:08:07.291 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-07T10:08:07.291 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-07T10:08:07.291 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:08:07.291 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-07T10:08:07.291 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-07T10:08:07.291 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-07T10:08:07.291 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-07T10:08:07.291 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-07T10:08:07.291 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-07T10:08:07.291 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "restful" 2026-03-07T10:08:07.291 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-07T10:08:07.291 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-07T10:08:07.292 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:08:07.292 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-07T10:08:07.292 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T10:08:07.292 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modified": "2026-03-07T10:07:53.158907+0000", 2026-03-07T10:08:07.292 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-07T10:08:07.292 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-07T10:08:07.292 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-07T10:08:07.292 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-07T10:08:07.292 INFO:teuthology.orchestra.run.vm02.stdout:mgr is available 2026-03-07T10:08:07.713 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-07T10:08:07.713 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [global] 2026-03-07T10:08:07.713 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout fsid = 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:08:07.713 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-07T10:08:07.713 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.102:3300,v1:192.168.123.102:6789] 2026-03-07T10:08:07.713 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-07T10:08:07.713 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-07T10:08:07.713 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-07T10:08:07.713 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-07T10:08:07.713 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-07T10:08:07.713 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-07T10:08:07.713 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-07T10:08:07.713 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-07T10:08:07.713 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [osd] 2026-03-07T10:08:07.713 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-07T10:08:07.713 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-07T10:08:07.713 INFO:teuthology.orchestra.run.vm02.stdout:Enabling cephadm module... 2026-03-07T10:08:07.866 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:07 vm02 bash[17011]: cluster 2026-03-07T10:08:06.578191+0000 mon.a (mon.0) 31 : cluster [DBG] mgrmap e3: a(active, since 1.04075s) 2026-03-07T10:08:07.866 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:07 vm02 bash[17011]: cluster 2026-03-07T10:08:06.578191+0000 mon.a (mon.0) 31 : cluster [DBG] mgrmap e3: a(active, since 1.04075s) 2026-03-07T10:08:07.866 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:07 vm02 bash[17011]: audit 2026-03-07T10:08:07.241995+0000 mon.a (mon.0) 32 : audit [DBG] from='client.? 192.168.123.102:0/3681381893' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-07T10:08:07.866 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:07 vm02 bash[17011]: audit 2026-03-07T10:08:07.241995+0000 mon.a (mon.0) 32 : audit [DBG] from='client.? 192.168.123.102:0/3681381893' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-07T10:08:08.961 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:08 vm02 bash[17284]: ignoring --setuser ceph since I am not root 2026-03-07T10:08:08.962 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:08 vm02 bash[17284]: ignoring --setgroup ceph since I am not root 2026-03-07T10:08:08.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:08 vm02 bash[17011]: cluster 2026-03-07T10:08:07.584783+0000 mon.a (mon.0) 33 : cluster [DBG] mgrmap e4: a(active, since 2s) 2026-03-07T10:08:08.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:08 vm02 bash[17011]: cluster 2026-03-07T10:08:07.584783+0000 mon.a (mon.0) 33 : cluster [DBG] mgrmap e4: a(active, since 2s) 2026-03-07T10:08:08.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:08 vm02 bash[17011]: audit 2026-03-07T10:08:07.651954+0000 mon.a (mon.0) 34 : audit [INF] from='client.? 192.168.123.102:0/3563255941' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-07T10:08:08.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:08 vm02 bash[17011]: audit 2026-03-07T10:08:07.651954+0000 mon.a (mon.0) 34 : audit [INF] from='client.? 192.168.123.102:0/3563255941' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-07T10:08:08.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:08 vm02 bash[17011]: audit 2026-03-07T10:08:07.655076+0000 mon.a (mon.0) 35 : audit [INF] from='client.? 192.168.123.102:0/3563255941' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-07T10:08:08.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:08 vm02 bash[17011]: audit 2026-03-07T10:08:07.655076+0000 mon.a (mon.0) 35 : audit [INF] from='client.? 192.168.123.102:0/3563255941' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-07T10:08:08.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:08 vm02 bash[17011]: audit 2026-03-07T10:08:08.083228+0000 mon.a (mon.0) 36 : audit [INF] from='client.? 192.168.123.102:0/1131728' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-07T10:08:08.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:08 vm02 bash[17011]: audit 2026-03-07T10:08:08.083228+0000 mon.a (mon.0) 36 : audit [INF] from='client.? 192.168.123.102:0/1131728' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-07T10:08:09.205 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-07T10:08:09.205 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 5, 2026-03-07T10:08:09.205 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-07T10:08:09.205 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "active_name": "a", 2026-03-07T10:08:09.205 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-07T10:08:09.205 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-07T10:08:09.205 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for the mgr to restart... 2026-03-07T10:08:09.205 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mgr epoch 5... 2026-03-07T10:08:09.287 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:08 vm02 bash[17284]: debug 2026-03-07T10:08:08.961+0000 7f7fd777f100 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-07T10:08:09.287 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:09 vm02 bash[17284]: debug 2026-03-07T10:08:09.089+0000 7f7fd777f100 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-07T10:08:09.961 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:09 vm02 bash[17284]: debug 2026-03-07T10:08:09.545+0000 7f7fd777f100 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-07T10:08:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:09 vm02 bash[17011]: audit 2026-03-07T10:08:08.656041+0000 mon.a (mon.0) 37 : audit [INF] from='client.? 192.168.123.102:0/1131728' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-07T10:08:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:09 vm02 bash[17011]: audit 2026-03-07T10:08:08.656041+0000 mon.a (mon.0) 37 : audit [INF] from='client.? 192.168.123.102:0/1131728' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-07T10:08:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:09 vm02 bash[17011]: cluster 2026-03-07T10:08:08.660360+0000 mon.a (mon.0) 38 : cluster [DBG] mgrmap e5: a(active, since 3s) 2026-03-07T10:08:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:09 vm02 bash[17011]: cluster 2026-03-07T10:08:08.660360+0000 mon.a (mon.0) 38 : cluster [DBG] mgrmap e5: a(active, since 3s) 2026-03-07T10:08:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:09 vm02 bash[17011]: audit 2026-03-07T10:08:09.120388+0000 mon.a (mon.0) 39 : audit [DBG] from='client.? 192.168.123.102:0/4182797360' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-07T10:08:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:09 vm02 bash[17011]: audit 2026-03-07T10:08:09.120388+0000 mon.a (mon.0) 39 : audit [DBG] from='client.? 192.168.123.102:0/4182797360' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-07T10:08:10.961 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:10 vm02 bash[17284]: debug 2026-03-07T10:08:10.697+0000 7f7fd777f100 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-07T10:08:12.408 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:12 vm02 bash[17284]: debug 2026-03-07T10:08:12.069+0000 7f7fd777f100 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-07T10:08:12.711 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:12 vm02 bash[17284]: debug 2026-03-07T10:08:12.401+0000 7f7fd777f100 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-07T10:08:13.179 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:12 vm02 bash[17284]: debug 2026-03-07T10:08:12.913+0000 7f7fd777f100 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-07T10:08:13.179 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:13 vm02 bash[17284]: debug 2026-03-07T10:08:13.045+0000 7f7fd777f100 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-07T10:08:13.462 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:13 vm02 bash[17284]: debug 2026-03-07T10:08:13.173+0000 7f7fd777f100 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-07T10:08:13.462 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:13 vm02 bash[17284]: debug 2026-03-07T10:08:13.333+0000 7f7fd777f100 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-07T10:08:13.961 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:13 vm02 bash[17284]: debug 2026-03-07T10:08:13.493+0000 7f7fd777f100 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-07T10:08:14.980 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:14 vm02 bash[17284]: debug 2026-03-07T10:08:14.721+0000 7f7fd777f100 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-07T10:08:14.980 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:14 vm02 bash[17284]: debug 2026-03-07T10:08:14.849+0000 7f7fd777f100 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-07T10:08:15.461 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:14 vm02 bash[17284]: debug 2026-03-07T10:08:14.973+0000 7f7fd777f100 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-07T10:08:15.787 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:15 vm02 bash[17284]: debug 2026-03-07T10:08:15.489+0000 7f7fd777f100 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-07T10:08:15.787 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:15 vm02 bash[17284]: debug 2026-03-07T10:08:15.637+0000 7f7fd777f100 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-07T10:08:16.139 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:15 vm02 bash[17284]: debug 2026-03-07T10:08:15.781+0000 7f7fd777f100 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-07T10:08:16.461 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:16 vm02 bash[17284]: debug 2026-03-07T10:08:16.133+0000 7f7fd777f100 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-07T10:08:16.961 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:16 vm02 bash[17284]: debug 2026-03-07T10:08:16.625+0000 7f7fd777f100 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-07T10:08:17.421 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:17 vm02 bash[17284]: debug 2026-03-07T10:08:17.153+0000 7f7fd777f100 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-07T10:08:17.421 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:17 vm02 bash[17284]: debug 2026-03-07T10:08:17.281+0000 7f7fd777f100 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-07T10:08:17.711 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:17 vm02 bash[17284]: debug 2026-03-07T10:08:17.417+0000 7f7fd777f100 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-07T10:08:18.211 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:17 vm02 bash[17284]: debug 2026-03-07T10:08:17.905+0000 7f7fd777f100 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-07T10:08:19.461 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:18 vm02 bash[17284]: debug 2026-03-07T10:08:18.961+0000 7f7fd777f100 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-07T10:08:19.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:19 vm02 bash[17011]: cluster 2026-03-07T10:08:18.965562+0000 mon.a (mon.0) 40 : cluster [INF] Active manager daemon a restarted 2026-03-07T10:08:19.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:19 vm02 bash[17011]: cluster 2026-03-07T10:08:18.965562+0000 mon.a (mon.0) 40 : cluster [INF] Active manager daemon a restarted 2026-03-07T10:08:19.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:19 vm02 bash[17011]: cluster 2026-03-07T10:08:18.965787+0000 mon.a (mon.0) 41 : cluster [INF] Activating manager daemon a 2026-03-07T10:08:19.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:19 vm02 bash[17011]: cluster 2026-03-07T10:08:18.965787+0000 mon.a (mon.0) 41 : cluster [INF] Activating manager daemon a 2026-03-07T10:08:19.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:19 vm02 bash[17011]: cluster 2026-03-07T10:08:18.973754+0000 mon.a (mon.0) 42 : cluster [DBG] osdmap e2: 0 total, 0 up, 0 in 2026-03-07T10:08:19.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:19 vm02 bash[17011]: cluster 2026-03-07T10:08:18.973754+0000 mon.a (mon.0) 42 : cluster [DBG] osdmap e2: 0 total, 0 up, 0 in 2026-03-07T10:08:19.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:19 vm02 bash[17011]: cluster 2026-03-07T10:08:18.973914+0000 mon.a (mon.0) 43 : cluster [DBG] mgrmap e6: a(active, starting, since 0.00823926s) 2026-03-07T10:08:19.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:19 vm02 bash[17011]: cluster 2026-03-07T10:08:18.973914+0000 mon.a (mon.0) 43 : cluster [DBG] mgrmap e6: a(active, starting, since 0.00823926s) 2026-03-07T10:08:19.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:19 vm02 bash[17011]: audit 2026-03-07T10:08:18.976689+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-07T10:08:19.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:19 vm02 bash[17011]: audit 2026-03-07T10:08:18.976689+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-07T10:08:19.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:19 vm02 bash[17011]: audit 2026-03-07T10:08:18.977520+0000 mon.a (mon.0) 45 : audit [DBG] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-07T10:08:19.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:19 vm02 bash[17011]: audit 2026-03-07T10:08:18.977520+0000 mon.a (mon.0) 45 : audit [DBG] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-07T10:08:19.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:19 vm02 bash[17011]: audit 2026-03-07T10:08:18.978415+0000 mon.a (mon.0) 46 : audit [DBG] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-07T10:08:19.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:19 vm02 bash[17011]: audit 2026-03-07T10:08:18.978415+0000 mon.a (mon.0) 46 : audit [DBG] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-07T10:08:19.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:19 vm02 bash[17011]: audit 2026-03-07T10:08:18.978473+0000 mon.a (mon.0) 47 : audit [DBG] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-07T10:08:19.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:19 vm02 bash[17011]: audit 2026-03-07T10:08:18.978473+0000 mon.a (mon.0) 47 : audit [DBG] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-07T10:08:19.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:19 vm02 bash[17011]: audit 2026-03-07T10:08:18.978807+0000 mon.a (mon.0) 48 : audit [DBG] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-07T10:08:19.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:19 vm02 bash[17011]: audit 2026-03-07T10:08:18.978807+0000 mon.a (mon.0) 48 : audit [DBG] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-07T10:08:19.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:19 vm02 bash[17011]: cluster 2026-03-07T10:08:18.984227+0000 mon.a (mon.0) 49 : cluster [INF] Manager daemon a is now available 2026-03-07T10:08:19.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:19 vm02 bash[17011]: cluster 2026-03-07T10:08:18.984227+0000 mon.a (mon.0) 49 : cluster [INF] Manager daemon a is now available 2026-03-07T10:08:19.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:19 vm02 bash[17011]: audit 2026-03-07T10:08:19.000470+0000 mon.a (mon.0) 50 : audit [INF] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' 2026-03-07T10:08:19.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:19 vm02 bash[17011]: audit 2026-03-07T10:08:19.000470+0000 mon.a (mon.0) 50 : audit [INF] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' 2026-03-07T10:08:19.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:19 vm02 bash[17011]: audit 2026-03-07T10:08:19.004152+0000 mon.a (mon.0) 51 : audit [INF] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' 2026-03-07T10:08:19.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:19 vm02 bash[17011]: audit 2026-03-07T10:08:19.004152+0000 mon.a (mon.0) 51 : audit [INF] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' 2026-03-07T10:08:20.033 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-07T10:08:20.033 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 7, 2026-03-07T10:08:20.033 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-07T10:08:20.033 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-07T10:08:20.033 INFO:teuthology.orchestra.run.vm02.stdout:mgr epoch 5 is available 2026-03-07T10:08:20.033 INFO:teuthology.orchestra.run.vm02.stdout:Setting orchestrator backend to cephadm... 2026-03-07T10:08:20.405 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:20 vm02 bash[17011]: cephadm 2026-03-07T10:08:18.997805+0000 mgr.a (mgr.14124) 1 : cephadm [INF] Found migration_current of "None". Setting to last migration. 2026-03-07T10:08:20.405 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:20 vm02 bash[17011]: cephadm 2026-03-07T10:08:18.997805+0000 mgr.a (mgr.14124) 1 : cephadm [INF] Found migration_current of "None". Setting to last migration. 2026-03-07T10:08:20.405 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:20 vm02 bash[17011]: audit 2026-03-07T10:08:19.021353+0000 mon.a (mon.0) 52 : audit [DBG] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:08:20.405 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:20 vm02 bash[17011]: audit 2026-03-07T10:08:19.021353+0000 mon.a (mon.0) 52 : audit [DBG] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:08:20.405 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:20 vm02 bash[17011]: audit 2026-03-07T10:08:19.026784+0000 mon.a (mon.0) 53 : audit [DBG] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:08:20.405 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:20 vm02 bash[17011]: audit 2026-03-07T10:08:19.026784+0000 mon.a (mon.0) 53 : audit [DBG] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:08:20.405 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:20 vm02 bash[17011]: audit 2026-03-07T10:08:19.031314+0000 mon.a (mon.0) 54 : audit [INF] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-07T10:08:20.405 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:20 vm02 bash[17011]: audit 2026-03-07T10:08:19.031314+0000 mon.a (mon.0) 54 : audit [INF] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-07T10:08:20.405 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:20 vm02 bash[17011]: audit 2026-03-07T10:08:19.038224+0000 mon.a (mon.0) 55 : audit [INF] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-07T10:08:20.405 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:20 vm02 bash[17011]: audit 2026-03-07T10:08:19.038224+0000 mon.a (mon.0) 55 : audit [INF] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-07T10:08:20.405 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:20 vm02 bash[17011]: audit 2026-03-07T10:08:19.266270+0000 mon.a (mon.0) 56 : audit [INF] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' 2026-03-07T10:08:20.406 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:20 vm02 bash[17011]: audit 2026-03-07T10:08:19.266270+0000 mon.a (mon.0) 56 : audit [INF] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' 2026-03-07T10:08:20.406 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:20 vm02 bash[17011]: audit 2026-03-07T10:08:19.269656+0000 mon.a (mon.0) 57 : audit [INF] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' 2026-03-07T10:08:20.406 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:20 vm02 bash[17011]: audit 2026-03-07T10:08:19.269656+0000 mon.a (mon.0) 57 : audit [INF] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' 2026-03-07T10:08:20.406 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:20 vm02 bash[17011]: cluster 2026-03-07T10:08:19.977150+0000 mon.a (mon.0) 58 : cluster [DBG] mgrmap e7: a(active, since 1.01148s) 2026-03-07T10:08:20.406 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:20 vm02 bash[17011]: cluster 2026-03-07T10:08:19.977150+0000 mon.a (mon.0) 58 : cluster [DBG] mgrmap e7: a(active, since 1.01148s) 2026-03-07T10:08:20.829 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout value unchanged 2026-03-07T10:08:20.829 INFO:teuthology.orchestra.run.vm02.stdout:Generating ssh key... 2026-03-07T10:08:21.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17011]: cephadm 2026-03-07T10:08:19.851997+0000 mgr.a (mgr.14124) 2 : cephadm [INF] [07/Mar/2026:10:08:19] ENGINE Bus STARTING 2026-03-07T10:08:21.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17011]: cephadm 2026-03-07T10:08:19.851997+0000 mgr.a (mgr.14124) 2 : cephadm [INF] [07/Mar/2026:10:08:19] ENGINE Bus STARTING 2026-03-07T10:08:21.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17011]: cephadm 2026-03-07T10:08:19.954378+0000 mgr.a (mgr.14124) 3 : cephadm [INF] [07/Mar/2026:10:08:19] ENGINE Serving on http://192.168.123.102:8765 2026-03-07T10:08:21.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17011]: cephadm 2026-03-07T10:08:19.954378+0000 mgr.a (mgr.14124) 3 : cephadm [INF] [07/Mar/2026:10:08:19] ENGINE Serving on http://192.168.123.102:8765 2026-03-07T10:08:21.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17011]: audit 2026-03-07T10:08:19.979731+0000 mgr.a (mgr.14124) 4 : audit [DBG] from='client.14128 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-07T10:08:21.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17011]: audit 2026-03-07T10:08:19.979731+0000 mgr.a (mgr.14124) 4 : audit [DBG] from='client.14128 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-07T10:08:21.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17011]: audit 2026-03-07T10:08:19.984462+0000 mgr.a (mgr.14124) 5 : audit [DBG] from='client.14128 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-07T10:08:21.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17011]: audit 2026-03-07T10:08:19.984462+0000 mgr.a (mgr.14124) 5 : audit [DBG] from='client.14128 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-07T10:08:21.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17011]: cephadm 2026-03-07T10:08:20.064222+0000 mgr.a (mgr.14124) 6 : cephadm [INF] [07/Mar/2026:10:08:20] ENGINE Serving on https://192.168.123.102:7150 2026-03-07T10:08:21.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17011]: cephadm 2026-03-07T10:08:20.064222+0000 mgr.a (mgr.14124) 6 : cephadm [INF] [07/Mar/2026:10:08:20] ENGINE Serving on https://192.168.123.102:7150 2026-03-07T10:08:21.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17011]: cephadm 2026-03-07T10:08:20.064307+0000 mgr.a (mgr.14124) 7 : cephadm [INF] [07/Mar/2026:10:08:20] ENGINE Bus STARTED 2026-03-07T10:08:21.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17011]: cephadm 2026-03-07T10:08:20.064307+0000 mgr.a (mgr.14124) 7 : cephadm [INF] [07/Mar/2026:10:08:20] ENGINE Bus STARTED 2026-03-07T10:08:21.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17011]: cephadm 2026-03-07T10:08:20.064701+0000 mgr.a (mgr.14124) 8 : cephadm [INF] [07/Mar/2026:10:08:20] ENGINE Client ('192.168.123.102', 54204) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-07T10:08:21.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17011]: cephadm 2026-03-07T10:08:20.064701+0000 mgr.a (mgr.14124) 8 : cephadm [INF] [07/Mar/2026:10:08:20] ENGINE Client ('192.168.123.102', 54204) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-07T10:08:21.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17011]: audit 2026-03-07T10:08:20.064817+0000 mon.a (mon.0) 59 : audit [DBG] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:08:21.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17011]: audit 2026-03-07T10:08:20.064817+0000 mon.a (mon.0) 59 : audit [DBG] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:08:21.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17011]: audit 2026-03-07T10:08:20.381766+0000 mgr.a (mgr.14124) 9 : audit [DBG] from='client.14136 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:08:21.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17011]: audit 2026-03-07T10:08:20.381766+0000 mgr.a (mgr.14124) 9 : audit [DBG] from='client.14136 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:08:21.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17011]: audit 2026-03-07T10:08:20.385813+0000 mon.a (mon.0) 60 : audit [INF] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' 2026-03-07T10:08:21.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17011]: audit 2026-03-07T10:08:20.385813+0000 mon.a (mon.0) 60 : audit [INF] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' 2026-03-07T10:08:21.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17011]: audit 2026-03-07T10:08:20.391317+0000 mon.a (mon.0) 61 : audit [DBG] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:08:21.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17011]: audit 2026-03-07T10:08:20.391317+0000 mon.a (mon.0) 61 : audit [DBG] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:08:21.675 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17284]: Generating public/private rsa key pair. 2026-03-07T10:08:21.675 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17284]: Your identification has been saved in /tmp/tmpl4mrccmq/key 2026-03-07T10:08:21.675 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17284]: Your public key has been saved in /tmp/tmpl4mrccmq/key.pub 2026-03-07T10:08:21.675 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17284]: The key fingerprint is: 2026-03-07T10:08:21.675 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17284]: SHA256:iMSe2jWzqnWZTR/2ajIW82wE4vkJr4i1Gf/hTJM/zl0 ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:08:21.675 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17284]: The key's randomart image is: 2026-03-07T10:08:21.675 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17284]: +---[RSA 3072]----+ 2026-03-07T10:08:21.675 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17284]: | | 2026-03-07T10:08:21.675 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17284]: | . | 2026-03-07T10:08:21.675 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17284]: | o | 2026-03-07T10:08:21.675 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17284]: | o o... | 2026-03-07T10:08:21.675 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17284]: | +.=oS.o | 2026-03-07T10:08:21.675 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17284]: | o .+Oo+.o | 2026-03-07T10:08:21.675 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17284]: | . = =+*B. . E | 2026-03-07T10:08:21.675 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17284]: | + O +B=*o . | 2026-03-07T10:08:21.675 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17284]: | o.= o++*=.. | 2026-03-07T10:08:21.675 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:21 vm02 bash[17284]: +----[SHA256]-----+ 2026-03-07T10:08:21.706 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDL/lIbyqWb7EsV3rPP0RxJyou5X09+3tos3a8YCDBRsqpn7OqCw/6+pCjMY9sXvIHBoWqH3Ygqz0cTTsmi7UPsj25NSHtIfZNjMF3eJB7GtzRaicrPWmA15GAvYjH2n+kWA7YH6TqkN/AlWYifSjQqWB9EJCI+rNLmAv3RfR3g1WsdrTTIfyqWmLX8puKKMNCVwv72+ljfdaABGlJKiJWq6QCKbr4lG42OFKIB+XBIw+9U5AaonK5hWzTXIxkiAsDIz34O4iiIdvMs5llBtQr1YOw/W05p9lBfHy/EUHMPGV6JOWJbEBEVMs5xh7W9HqnEpvIIJsRhCsUZULgbyKId4KNTQRpnaLrB45HBTVE/FNt3Ag8BiK2ViwbPUVTvdfAUS/ZGBN1KmEwRV4FAZVTuC6VHHnysFBYjPfHxWpou2tbp9NvP8UoYwe449rQ87h0yVwBjQsNCP2FM07+HxbvrA6EqoztzSp7nnMQWAc/0kHHL2qaz6s3R70FsTOKH1hM= ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:08:21.707 INFO:teuthology.orchestra.run.vm02.stdout:Wrote public SSH key to /home/ubuntu/cephtest/ceph.pub 2026-03-07T10:08:21.707 INFO:teuthology.orchestra.run.vm02.stdout:Adding key to root@localhost authorized_keys... 2026-03-07T10:08:21.707 INFO:teuthology.orchestra.run.vm02.stdout:Adding host vm02... 2026-03-07T10:08:22.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:22 vm02 bash[17011]: audit 2026-03-07T10:08:20.783341+0000 mgr.a (mgr.14124) 10 : audit [DBG] from='client.14138 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:08:22.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:22 vm02 bash[17011]: audit 2026-03-07T10:08:20.783341+0000 mgr.a (mgr.14124) 10 : audit [DBG] from='client.14138 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:08:22.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:22 vm02 bash[17011]: audit 2026-03-07T10:08:21.172206+0000 mgr.a (mgr.14124) 11 : audit [DBG] from='client.14140 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:08:22.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:22 vm02 bash[17011]: audit 2026-03-07T10:08:21.172206+0000 mgr.a (mgr.14124) 11 : audit [DBG] from='client.14140 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:08:22.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:22 vm02 bash[17011]: cephadm 2026-03-07T10:08:21.172455+0000 mgr.a (mgr.14124) 12 : cephadm [INF] Generating ssh key... 2026-03-07T10:08:22.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:22 vm02 bash[17011]: cephadm 2026-03-07T10:08:21.172455+0000 mgr.a (mgr.14124) 12 : cephadm [INF] Generating ssh key... 2026-03-07T10:08:22.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:22 vm02 bash[17011]: audit 2026-03-07T10:08:21.229675+0000 mon.a (mon.0) 62 : audit [INF] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' 2026-03-07T10:08:22.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:22 vm02 bash[17011]: audit 2026-03-07T10:08:21.229675+0000 mon.a (mon.0) 62 : audit [INF] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' 2026-03-07T10:08:22.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:22 vm02 bash[17011]: audit 2026-03-07T10:08:21.232078+0000 mon.a (mon.0) 63 : audit [INF] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' 2026-03-07T10:08:22.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:22 vm02 bash[17011]: audit 2026-03-07T10:08:21.232078+0000 mon.a (mon.0) 63 : audit [INF] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' 2026-03-07T10:08:22.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:22 vm02 bash[17011]: cluster 2026-03-07T10:08:21.389664+0000 mon.a (mon.0) 64 : cluster [DBG] mgrmap e8: a(active, since 2s) 2026-03-07T10:08:22.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:22 vm02 bash[17011]: cluster 2026-03-07T10:08:21.389664+0000 mon.a (mon.0) 64 : cluster [DBG] mgrmap e8: a(active, since 2s) 2026-03-07T10:08:22.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:22 vm02 bash[17011]: audit 2026-03-07T10:08:21.659640+0000 mgr.a (mgr.14124) 13 : audit [DBG] from='client.14142 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:08:22.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:22 vm02 bash[17011]: audit 2026-03-07T10:08:21.659640+0000 mgr.a (mgr.14124) 13 : audit [DBG] from='client.14142 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:08:23.404 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:23 vm02 bash[17011]: audit 2026-03-07T10:08:22.031321+0000 mgr.a (mgr.14124) 14 : audit [DBG] from='client.14144 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm02", "addr": "192.168.123.102", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:08:23.404 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:23 vm02 bash[17011]: audit 2026-03-07T10:08:22.031321+0000 mgr.a (mgr.14124) 14 : audit [DBG] from='client.14144 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm02", "addr": "192.168.123.102", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:08:24.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:24 vm02 bash[17011]: cephadm 2026-03-07T10:08:22.916612+0000 mgr.a (mgr.14124) 15 : cephadm [INF] Deploying cephadm binary to vm02 2026-03-07T10:08:24.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:24 vm02 bash[17011]: cephadm 2026-03-07T10:08:22.916612+0000 mgr.a (mgr.14124) 15 : cephadm [INF] Deploying cephadm binary to vm02 2026-03-07T10:08:24.753 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout Added host 'vm02' with addr '192.168.123.102' 2026-03-07T10:08:24.753 INFO:teuthology.orchestra.run.vm02.stdout:Deploying unmanaged mon service... 2026-03-07T10:08:25.171 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout Scheduled mon update... 2026-03-07T10:08:25.171 INFO:teuthology.orchestra.run.vm02.stdout:Deploying unmanaged mgr service... 2026-03-07T10:08:25.576 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout Scheduled mgr update... 2026-03-07T10:08:25.793 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:25 vm02 bash[17011]: audit 2026-03-07T10:08:24.696380+0000 mon.a (mon.0) 65 : audit [INF] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' 2026-03-07T10:08:25.793 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:25 vm02 bash[17011]: audit 2026-03-07T10:08:24.696380+0000 mon.a (mon.0) 65 : audit [INF] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' 2026-03-07T10:08:25.793 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:25 vm02 bash[17011]: cephadm 2026-03-07T10:08:24.696835+0000 mgr.a (mgr.14124) 16 : cephadm [INF] Added host vm02 2026-03-07T10:08:25.793 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:25 vm02 bash[17011]: cephadm 2026-03-07T10:08:24.696835+0000 mgr.a (mgr.14124) 16 : cephadm [INF] Added host vm02 2026-03-07T10:08:25.793 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:25 vm02 bash[17011]: audit 2026-03-07T10:08:24.697174+0000 mon.a (mon.0) 66 : audit [DBG] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:08:25.793 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:25 vm02 bash[17011]: audit 2026-03-07T10:08:24.697174+0000 mon.a (mon.0) 66 : audit [DBG] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:08:25.793 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:25 vm02 bash[17011]: audit 2026-03-07T10:08:25.122314+0000 mon.a (mon.0) 67 : audit [INF] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' 2026-03-07T10:08:25.793 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:25 vm02 bash[17011]: audit 2026-03-07T10:08:25.122314+0000 mon.a (mon.0) 67 : audit [INF] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' 2026-03-07T10:08:25.793 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:25 vm02 bash[17011]: audit 2026-03-07T10:08:25.519241+0000 mon.a (mon.0) 68 : audit [INF] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' 2026-03-07T10:08:25.793 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:25 vm02 bash[17011]: audit 2026-03-07T10:08:25.519241+0000 mon.a (mon.0) 68 : audit [INF] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' 2026-03-07T10:08:26.668 INFO:teuthology.orchestra.run.vm02.stdout:Enabling the dashboard module... 2026-03-07T10:08:26.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:26 vm02 bash[17011]: audit 2026-03-07T10:08:25.118428+0000 mgr.a (mgr.14124) 17 : audit [DBG] from='client.14146 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:08:26.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:26 vm02 bash[17011]: audit 2026-03-07T10:08:25.118428+0000 mgr.a (mgr.14124) 17 : audit [DBG] from='client.14146 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:08:26.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:26 vm02 bash[17011]: cephadm 2026-03-07T10:08:25.119471+0000 mgr.a (mgr.14124) 18 : cephadm [INF] Saving service mon spec with placement count:5 2026-03-07T10:08:26.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:26 vm02 bash[17011]: cephadm 2026-03-07T10:08:25.119471+0000 mgr.a (mgr.14124) 18 : cephadm [INF] Saving service mon spec with placement count:5 2026-03-07T10:08:26.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:26 vm02 bash[17011]: audit 2026-03-07T10:08:25.515989+0000 mgr.a (mgr.14124) 19 : audit [DBG] from='client.14148 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:08:26.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:26 vm02 bash[17011]: audit 2026-03-07T10:08:25.515989+0000 mgr.a (mgr.14124) 19 : audit [DBG] from='client.14148 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:08:26.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:26 vm02 bash[17011]: cephadm 2026-03-07T10:08:25.516793+0000 mgr.a (mgr.14124) 20 : cephadm [INF] Saving service mgr spec with placement count:2 2026-03-07T10:08:26.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:26 vm02 bash[17011]: cephadm 2026-03-07T10:08:25.516793+0000 mgr.a (mgr.14124) 20 : cephadm [INF] Saving service mgr spec with placement count:2 2026-03-07T10:08:26.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:26 vm02 bash[17011]: audit 2026-03-07T10:08:25.833947+0000 mon.a (mon.0) 69 : audit [INF] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' 2026-03-07T10:08:26.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:26 vm02 bash[17011]: audit 2026-03-07T10:08:25.833947+0000 mon.a (mon.0) 69 : audit [INF] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' 2026-03-07T10:08:26.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:26 vm02 bash[17011]: audit 2026-03-07T10:08:25.975624+0000 mon.a (mon.0) 70 : audit [INF] from='client.? 192.168.123.102:0/1556816821' entity='client.admin' 2026-03-07T10:08:26.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:26 vm02 bash[17011]: audit 2026-03-07T10:08:25.975624+0000 mon.a (mon.0) 70 : audit [INF] from='client.? 192.168.123.102:0/1556816821' entity='client.admin' 2026-03-07T10:08:26.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:26 vm02 bash[17011]: audit 2026-03-07T10:08:26.180956+0000 mon.a (mon.0) 71 : audit [INF] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' 2026-03-07T10:08:26.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:26 vm02 bash[17011]: audit 2026-03-07T10:08:26.180956+0000 mon.a (mon.0) 71 : audit [INF] from='mgr.14124 192.168.123.102:0/1845649563' entity='mgr.a' 2026-03-07T10:08:26.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:26 vm02 bash[17011]: audit 2026-03-07T10:08:26.619300+0000 mon.a (mon.0) 72 : audit [INF] from='client.? 192.168.123.102:0/2115961514' entity='client.admin' 2026-03-07T10:08:26.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:26 vm02 bash[17011]: audit 2026-03-07T10:08:26.619300+0000 mon.a (mon.0) 72 : audit [INF] from='client.? 192.168.123.102:0/2115961514' entity='client.admin' 2026-03-07T10:08:28.062 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:27 vm02 bash[17284]: ignoring --setuser ceph since I am not root 2026-03-07T10:08:28.062 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:27 vm02 bash[17284]: ignoring --setgroup ceph since I am not root 2026-03-07T10:08:28.062 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:27 vm02 bash[17011]: audit 2026-03-07T10:08:27.029024+0000 mon.a (mon.0) 73 : audit [INF] from='client.? 192.168.123.102:0/654036003' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-07T10:08:28.062 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:27 vm02 bash[17011]: audit 2026-03-07T10:08:27.029024+0000 mon.a (mon.0) 73 : audit [INF] from='client.? 192.168.123.102:0/654036003' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-07T10:08:28.360 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:28 vm02 bash[17284]: debug 2026-03-07T10:08:28.221+0000 7f9aa7621100 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-07T10:08:28.459 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-07T10:08:28.459 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 9, 2026-03-07T10:08:28.459 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-07T10:08:28.459 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "active_name": "a", 2026-03-07T10:08:28.459 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-07T10:08:28.459 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-07T10:08:28.459 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for the mgr to restart... 2026-03-07T10:08:28.460 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mgr epoch 9... 2026-03-07T10:08:28.712 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:28 vm02 bash[17284]: debug 2026-03-07T10:08:28.353+0000 7f9aa7621100 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-07T10:08:29.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:28 vm02 bash[17011]: audit 2026-03-07T10:08:27.905470+0000 mon.a (mon.0) 74 : audit [INF] from='client.? 192.168.123.102:0/654036003' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-07T10:08:29.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:28 vm02 bash[17011]: audit 2026-03-07T10:08:27.905470+0000 mon.a (mon.0) 74 : audit [INF] from='client.? 192.168.123.102:0/654036003' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-07T10:08:29.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:28 vm02 bash[17011]: cluster 2026-03-07T10:08:27.910874+0000 mon.a (mon.0) 75 : cluster [DBG] mgrmap e9: a(active, since 8s) 2026-03-07T10:08:29.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:28 vm02 bash[17011]: cluster 2026-03-07T10:08:27.910874+0000 mon.a (mon.0) 75 : cluster [DBG] mgrmap e9: a(active, since 8s) 2026-03-07T10:08:29.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:28 vm02 bash[17011]: audit 2026-03-07T10:08:28.404862+0000 mon.a (mon.0) 76 : audit [DBG] from='client.? 192.168.123.102:0/2808486111' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-07T10:08:29.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:28 vm02 bash[17011]: audit 2026-03-07T10:08:28.404862+0000 mon.a (mon.0) 76 : audit [DBG] from='client.? 192.168.123.102:0/2808486111' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-07T10:08:29.212 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:28 vm02 bash[17284]: debug 2026-03-07T10:08:28.793+0000 7f9aa7621100 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-07T10:08:30.462 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:29 vm02 bash[17284]: debug 2026-03-07T10:08:29.985+0000 7f9aa7621100 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-07T10:08:31.566 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:31 vm02 bash[17284]: debug 2026-03-07T10:08:31.317+0000 7f9aa7621100 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-07T10:08:31.931 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:31 vm02 bash[17284]: debug 2026-03-07T10:08:31.561+0000 7f9aa7621100 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-07T10:08:32.212 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:31 vm02 bash[17284]: debug 2026-03-07T10:08:31.925+0000 7f9aa7621100 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-07T10:08:32.212 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:31 vm02 bash[17284]: debug 2026-03-07T10:08:31.993+0000 7f9aa7621100 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-07T10:08:32.212 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:32 vm02 bash[17284]: debug 2026-03-07T10:08:32.057+0000 7f9aa7621100 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-07T10:08:32.212 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:32 vm02 bash[17284]: debug 2026-03-07T10:08:32.141+0000 7f9aa7621100 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-07T10:08:32.712 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:32 vm02 bash[17284]: debug 2026-03-07T10:08:32.233+0000 7f9aa7621100 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-07T10:08:33.712 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:33 vm02 bash[17284]: debug 2026-03-07T10:08:33.249+0000 7f9aa7621100 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-07T10:08:33.712 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:33 vm02 bash[17284]: debug 2026-03-07T10:08:33.313+0000 7f9aa7621100 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-07T10:08:33.712 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:33 vm02 bash[17284]: debug 2026-03-07T10:08:33.381+0000 7f9aa7621100 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-07T10:08:34.145 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:33 vm02 bash[17284]: debug 2026-03-07T10:08:33.745+0000 7f9aa7621100 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-07T10:08:34.145 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:33 vm02 bash[17284]: debug 2026-03-07T10:08:33.809+0000 7f9aa7621100 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-07T10:08:34.145 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:33 vm02 bash[17284]: debug 2026-03-07T10:08:33.877+0000 7f9aa7621100 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-07T10:08:34.145 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:34 vm02 bash[17284]: debug 2026-03-07T10:08:34.137+0000 7f9aa7621100 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-07T10:08:34.906 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:34 vm02 bash[17284]: debug 2026-03-07T10:08:34.465+0000 7f9aa7621100 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-07T10:08:35.211 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:34 vm02 bash[17284]: debug 2026-03-07T10:08:34.901+0000 7f9aa7621100 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-07T10:08:35.212 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:34 vm02 bash[17284]: debug 2026-03-07T10:08:34.961+0000 7f9aa7621100 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-07T10:08:35.212 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:35 vm02 bash[17284]: debug 2026-03-07T10:08:35.029+0000 7f9aa7621100 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-07T10:08:35.712 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:35 vm02 bash[17284]: debug 2026-03-07T10:08:35.353+0000 7f9aa7621100 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-07T10:08:36.712 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:08:36 vm02 bash[17284]: debug 2026-03-07T10:08:36.249+0000 7f9aa7621100 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-07T10:08:36.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:36 vm02 bash[17011]: cluster 2026-03-07T10:08:36.254020+0000 mon.a (mon.0) 77 : cluster [INF] Active manager daemon a restarted 2026-03-07T10:08:36.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:36 vm02 bash[17011]: cluster 2026-03-07T10:08:36.254020+0000 mon.a (mon.0) 77 : cluster [INF] Active manager daemon a restarted 2026-03-07T10:08:36.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:36 vm02 bash[17011]: cluster 2026-03-07T10:08:36.254607+0000 mon.a (mon.0) 78 : cluster [INF] Activating manager daemon a 2026-03-07T10:08:36.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:36 vm02 bash[17011]: cluster 2026-03-07T10:08:36.254607+0000 mon.a (mon.0) 78 : cluster [INF] Activating manager daemon a 2026-03-07T10:08:36.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:36 vm02 bash[17011]: cluster 2026-03-07T10:08:36.262657+0000 mon.a (mon.0) 79 : cluster [DBG] osdmap e3: 0 total, 0 up, 0 in 2026-03-07T10:08:36.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:36 vm02 bash[17011]: cluster 2026-03-07T10:08:36.262657+0000 mon.a (mon.0) 79 : cluster [DBG] osdmap e3: 0 total, 0 up, 0 in 2026-03-07T10:08:36.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:36 vm02 bash[17011]: cluster 2026-03-07T10:08:36.262754+0000 mon.a (mon.0) 80 : cluster [DBG] mgrmap e10: a(active, starting, since 0.00837469s) 2026-03-07T10:08:36.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:36 vm02 bash[17011]: cluster 2026-03-07T10:08:36.262754+0000 mon.a (mon.0) 80 : cluster [DBG] mgrmap e10: a(active, starting, since 0.00837469s) 2026-03-07T10:08:36.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:36 vm02 bash[17011]: audit 2026-03-07T10:08:36.265302+0000 mon.a (mon.0) 81 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-07T10:08:36.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:36 vm02 bash[17011]: audit 2026-03-07T10:08:36.265302+0000 mon.a (mon.0) 81 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-07T10:08:36.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:36 vm02 bash[17011]: audit 2026-03-07T10:08:36.265798+0000 mon.a (mon.0) 82 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-07T10:08:36.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:36 vm02 bash[17011]: audit 2026-03-07T10:08:36.265798+0000 mon.a (mon.0) 82 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-07T10:08:36.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:36 vm02 bash[17011]: audit 2026-03-07T10:08:36.267701+0000 mon.a (mon.0) 83 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-07T10:08:36.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:36 vm02 bash[17011]: audit 2026-03-07T10:08:36.267701+0000 mon.a (mon.0) 83 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-07T10:08:36.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:36 vm02 bash[17011]: audit 2026-03-07T10:08:36.268232+0000 mon.a (mon.0) 84 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-07T10:08:36.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:36 vm02 bash[17011]: audit 2026-03-07T10:08:36.268232+0000 mon.a (mon.0) 84 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-07T10:08:36.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:36 vm02 bash[17011]: audit 2026-03-07T10:08:36.268691+0000 mon.a (mon.0) 85 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-07T10:08:36.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:36 vm02 bash[17011]: audit 2026-03-07T10:08:36.268691+0000 mon.a (mon.0) 85 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-07T10:08:36.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:36 vm02 bash[17011]: cluster 2026-03-07T10:08:36.274413+0000 mon.a (mon.0) 86 : cluster [INF] Manager daemon a is now available 2026-03-07T10:08:36.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:36 vm02 bash[17011]: cluster 2026-03-07T10:08:36.274413+0000 mon.a (mon.0) 86 : cluster [INF] Manager daemon a is now available 2026-03-07T10:08:36.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:36 vm02 bash[17011]: audit 2026-03-07T10:08:36.296429+0000 mon.a (mon.0) 87 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:08:36.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:36 vm02 bash[17011]: audit 2026-03-07T10:08:36.296429+0000 mon.a (mon.0) 87 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:08:36.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:36 vm02 bash[17011]: audit 2026-03-07T10:08:36.309329+0000 mon.a (mon.0) 88 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-07T10:08:36.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:36 vm02 bash[17011]: audit 2026-03-07T10:08:36.309329+0000 mon.a (mon.0) 88 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-07T10:08:37.341 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-07T10:08:37.341 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 11, 2026-03-07T10:08:37.341 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-07T10:08:37.341 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-07T10:08:37.341 INFO:teuthology.orchestra.run.vm02.stdout:mgr epoch 9 is available 2026-03-07T10:08:37.341 INFO:teuthology.orchestra.run.vm02.stdout:Generating a dashboard self-signed certificate... 2026-03-07T10:08:37.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:37 vm02 bash[17011]: audit 2026-03-07T10:08:36.334469+0000 mon.a (mon.0) 89 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-07T10:08:37.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:37 vm02 bash[17011]: audit 2026-03-07T10:08:36.334469+0000 mon.a (mon.0) 89 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-07T10:08:37.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:37 vm02 bash[17011]: cluster 2026-03-07T10:08:37.266756+0000 mon.a (mon.0) 90 : cluster [DBG] mgrmap e11: a(active, since 1.01238s) 2026-03-07T10:08:37.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:37 vm02 bash[17011]: cluster 2026-03-07T10:08:37.266756+0000 mon.a (mon.0) 90 : cluster [DBG] mgrmap e11: a(active, since 1.01238s) 2026-03-07T10:08:37.788 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout Self-signed certificate created 2026-03-07T10:08:37.789 INFO:teuthology.orchestra.run.vm02.stdout:Creating initial admin user... 2026-03-07T10:08:38.342 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout {"username": "admin", "password": "$2b$12$yZm5XzbfeGoTiPnVQq7njOQuSUp.6Ogh737HjkjOGT4DrcocoFqcK", "roles": ["administrator"], "name": null, "email": null, "lastUpdate": 1772878118, "enabled": true, "pwdExpirationDate": null, "pwdUpdateRequired": true} 2026-03-07T10:08:38.342 INFO:teuthology.orchestra.run.vm02.stdout:Fetching dashboard port number... 2026-03-07T10:08:38.752 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 8443 2026-03-07T10:08:38.752 INFO:teuthology.orchestra.run.vm02.stdout:firewalld does not appear to be present 2026-03-07T10:08:38.752 INFO:teuthology.orchestra.run.vm02.stdout:Not possible to open ports <[8443]>. firewalld.service is not available 2026-03-07T10:08:38.753 INFO:teuthology.orchestra.run.vm02.stdout:Ceph Dashboard is now available at: 2026-03-07T10:08:38.753 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-07T10:08:38.753 INFO:teuthology.orchestra.run.vm02.stdout: URL: https://vm02.local:8443/ 2026-03-07T10:08:38.753 INFO:teuthology.orchestra.run.vm02.stdout: User: admin 2026-03-07T10:08:38.753 INFO:teuthology.orchestra.run.vm02.stdout: Password: 2l7nghkh0n 2026-03-07T10:08:38.753 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-07T10:08:38.753 INFO:teuthology.orchestra.run.vm02.stdout:Saving cluster configuration to /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/config directory 2026-03-07T10:08:38.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:38 vm02 bash[17011]: cephadm 2026-03-07T10:08:37.014341+0000 mgr.a (mgr.14156) 1 : cephadm [INF] [07/Mar/2026:10:08:37] ENGINE Bus STARTING 2026-03-07T10:08:38.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:38 vm02 bash[17011]: cephadm 2026-03-07T10:08:37.014341+0000 mgr.a (mgr.14156) 1 : cephadm [INF] [07/Mar/2026:10:08:37] ENGINE Bus STARTING 2026-03-07T10:08:38.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:38 vm02 bash[17011]: cephadm 2026-03-07T10:08:37.117731+0000 mgr.a (mgr.14156) 2 : cephadm [INF] [07/Mar/2026:10:08:37] ENGINE Serving on http://192.168.123.102:8765 2026-03-07T10:08:38.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:38 vm02 bash[17011]: cephadm 2026-03-07T10:08:37.117731+0000 mgr.a (mgr.14156) 2 : cephadm [INF] [07/Mar/2026:10:08:37] ENGINE Serving on http://192.168.123.102:8765 2026-03-07T10:08:38.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:38 vm02 bash[17011]: cephadm 2026-03-07T10:08:37.228400+0000 mgr.a (mgr.14156) 3 : cephadm [INF] [07/Mar/2026:10:08:37] ENGINE Serving on https://192.168.123.102:7150 2026-03-07T10:08:38.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:38 vm02 bash[17011]: cephadm 2026-03-07T10:08:37.228400+0000 mgr.a (mgr.14156) 3 : cephadm [INF] [07/Mar/2026:10:08:37] ENGINE Serving on https://192.168.123.102:7150 2026-03-07T10:08:38.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:38 vm02 bash[17011]: cephadm 2026-03-07T10:08:37.228454+0000 mgr.a (mgr.14156) 4 : cephadm [INF] [07/Mar/2026:10:08:37] ENGINE Bus STARTED 2026-03-07T10:08:38.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:38 vm02 bash[17011]: cephadm 2026-03-07T10:08:37.228454+0000 mgr.a (mgr.14156) 4 : cephadm [INF] [07/Mar/2026:10:08:37] ENGINE Bus STARTED 2026-03-07T10:08:38.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:38 vm02 bash[17011]: cephadm 2026-03-07T10:08:37.228864+0000 mgr.a (mgr.14156) 5 : cephadm [INF] [07/Mar/2026:10:08:37] ENGINE Client ('192.168.123.102', 48328) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-07T10:08:38.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:38 vm02 bash[17011]: cephadm 2026-03-07T10:08:37.228864+0000 mgr.a (mgr.14156) 5 : cephadm [INF] [07/Mar/2026:10:08:37] ENGINE Client ('192.168.123.102', 48328) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-07T10:08:38.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:38 vm02 bash[17011]: audit 2026-03-07T10:08:37.269272+0000 mgr.a (mgr.14156) 6 : audit [DBG] from='client.14160 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-07T10:08:38.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:38 vm02 bash[17011]: audit 2026-03-07T10:08:37.269272+0000 mgr.a (mgr.14156) 6 : audit [DBG] from='client.14160 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-07T10:08:38.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:38 vm02 bash[17011]: audit 2026-03-07T10:08:37.273281+0000 mgr.a (mgr.14156) 7 : audit [DBG] from='client.14160 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-07T10:08:38.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:38 vm02 bash[17011]: audit 2026-03-07T10:08:37.273281+0000 mgr.a (mgr.14156) 7 : audit [DBG] from='client.14160 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-07T10:08:38.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:38 vm02 bash[17011]: audit 2026-03-07T10:08:37.677608+0000 mgr.a (mgr.14156) 8 : audit [DBG] from='client.14168 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:08:38.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:38 vm02 bash[17011]: audit 2026-03-07T10:08:37.677608+0000 mgr.a (mgr.14156) 8 : audit [DBG] from='client.14168 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:08:38.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:38 vm02 bash[17011]: audit 2026-03-07T10:08:37.735160+0000 mon.a (mon.0) 91 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:38.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:38 vm02 bash[17011]: audit 2026-03-07T10:08:37.735160+0000 mon.a (mon.0) 91 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:38.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:38 vm02 bash[17011]: audit 2026-03-07T10:08:37.739026+0000 mon.a (mon.0) 92 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:38.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:38 vm02 bash[17011]: audit 2026-03-07T10:08:37.739026+0000 mon.a (mon.0) 92 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:38.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:38 vm02 bash[17011]: audit 2026-03-07T10:08:38.291063+0000 mon.a (mon.0) 93 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:38.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:38 vm02 bash[17011]: audit 2026-03-07T10:08:38.291063+0000 mon.a (mon.0) 93 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:38.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:38 vm02 bash[17011]: audit 2026-03-07T10:08:38.683244+0000 mon.a (mon.0) 94 : audit [DBG] from='client.? 192.168.123.102:0/1426769952' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-07T10:08:38.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:38 vm02 bash[17011]: audit 2026-03-07T10:08:38.683244+0000 mon.a (mon.0) 94 : audit [DBG] from='client.? 192.168.123.102:0/1426769952' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-07T10:08:39.306 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stderr set mgr/dashboard/cluster/status 2026-03-07T10:08:39.306 INFO:teuthology.orchestra.run.vm02.stdout:You can access the Ceph CLI as following in case of multi-cluster or non-default config: 2026-03-07T10:08:39.306 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-07T10:08:39.306 INFO:teuthology.orchestra.run.vm02.stdout: sudo /home/ubuntu/cephtest/cephadm shell --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring 2026-03-07T10:08:39.306 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-07T10:08:39.306 INFO:teuthology.orchestra.run.vm02.stdout:Or, if you are only running a single cluster on this host: 2026-03-07T10:08:39.306 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-07T10:08:39.306 INFO:teuthology.orchestra.run.vm02.stdout: sudo /home/ubuntu/cephtest/cephadm shell 2026-03-07T10:08:39.306 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-07T10:08:39.306 INFO:teuthology.orchestra.run.vm02.stdout:Please consider enabling telemetry to help improve Ceph: 2026-03-07T10:08:39.306 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-07T10:08:39.307 INFO:teuthology.orchestra.run.vm02.stdout: ceph telemetry on 2026-03-07T10:08:39.307 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-07T10:08:39.307 INFO:teuthology.orchestra.run.vm02.stdout:For more information see: 2026-03-07T10:08:39.307 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-07T10:08:39.307 INFO:teuthology.orchestra.run.vm02.stdout: https://docs.ceph.com/en/latest/mgr/telemetry/ 2026-03-07T10:08:39.307 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-07T10:08:39.307 INFO:teuthology.orchestra.run.vm02.stdout:Bootstrap complete. 2026-03-07T10:08:39.336 INFO:tasks.cephadm:Fetching config... 2026-03-07T10:08:39.336 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-07T10:08:39.336 DEBUG:teuthology.orchestra.run.vm02:> dd if=/etc/ceph/ceph.conf of=/dev/stdout 2026-03-07T10:08:39.339 INFO:tasks.cephadm:Fetching client.admin keyring... 2026-03-07T10:08:39.339 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-07T10:08:39.339 DEBUG:teuthology.orchestra.run.vm02:> dd if=/etc/ceph/ceph.client.admin.keyring of=/dev/stdout 2026-03-07T10:08:39.383 INFO:tasks.cephadm:Fetching mon keyring... 2026-03-07T10:08:39.383 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-07T10:08:39.383 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/keyring of=/dev/stdout 2026-03-07T10:08:39.431 INFO:tasks.cephadm:Fetching pub ssh key... 2026-03-07T10:08:39.431 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-07T10:08:39.431 DEBUG:teuthology.orchestra.run.vm02:> dd if=/home/ubuntu/cephtest/ceph.pub of=/dev/stdout 2026-03-07T10:08:39.475 INFO:tasks.cephadm:Installing pub ssh key for root users... 2026-03-07T10:08:39.475 DEBUG:teuthology.orchestra.run.vm02:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDL/lIbyqWb7EsV3rPP0RxJyou5X09+3tos3a8YCDBRsqpn7OqCw/6+pCjMY9sXvIHBoWqH3Ygqz0cTTsmi7UPsj25NSHtIfZNjMF3eJB7GtzRaicrPWmA15GAvYjH2n+kWA7YH6TqkN/AlWYifSjQqWB9EJCI+rNLmAv3RfR3g1WsdrTTIfyqWmLX8puKKMNCVwv72+ljfdaABGlJKiJWq6QCKbr4lG42OFKIB+XBIw+9U5AaonK5hWzTXIxkiAsDIz34O4iiIdvMs5llBtQr1YOw/W05p9lBfHy/EUHMPGV6JOWJbEBEVMs5xh7W9HqnEpvIIJsRhCsUZULgbyKId4KNTQRpnaLrB45HBTVE/FNt3Ag8BiK2ViwbPUVTvdfAUS/ZGBN1KmEwRV4FAZVTuC6VHHnysFBYjPfHxWpou2tbp9NvP8UoYwe449rQ87h0yVwBjQsNCP2FM07+HxbvrA6EqoztzSp7nnMQWAc/0kHHL2qaz6s3R70FsTOKH1hM= ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-07T10:08:39.530 INFO:teuthology.orchestra.run.vm02.stdout:ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDL/lIbyqWb7EsV3rPP0RxJyou5X09+3tos3a8YCDBRsqpn7OqCw/6+pCjMY9sXvIHBoWqH3Ygqz0cTTsmi7UPsj25NSHtIfZNjMF3eJB7GtzRaicrPWmA15GAvYjH2n+kWA7YH6TqkN/AlWYifSjQqWB9EJCI+rNLmAv3RfR3g1WsdrTTIfyqWmLX8puKKMNCVwv72+ljfdaABGlJKiJWq6QCKbr4lG42OFKIB+XBIw+9U5AaonK5hWzTXIxkiAsDIz34O4iiIdvMs5llBtQr1YOw/W05p9lBfHy/EUHMPGV6JOWJbEBEVMs5xh7W9HqnEpvIIJsRhCsUZULgbyKId4KNTQRpnaLrB45HBTVE/FNt3Ag8BiK2ViwbPUVTvdfAUS/ZGBN1KmEwRV4FAZVTuC6VHHnysFBYjPfHxWpou2tbp9NvP8UoYwe449rQ87h0yVwBjQsNCP2FM07+HxbvrA6EqoztzSp7nnMQWAc/0kHHL2qaz6s3R70FsTOKH1hM= ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:08:39.535 DEBUG:teuthology.orchestra.run.vm10:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDL/lIbyqWb7EsV3rPP0RxJyou5X09+3tos3a8YCDBRsqpn7OqCw/6+pCjMY9sXvIHBoWqH3Ygqz0cTTsmi7UPsj25NSHtIfZNjMF3eJB7GtzRaicrPWmA15GAvYjH2n+kWA7YH6TqkN/AlWYifSjQqWB9EJCI+rNLmAv3RfR3g1WsdrTTIfyqWmLX8puKKMNCVwv72+ljfdaABGlJKiJWq6QCKbr4lG42OFKIB+XBIw+9U5AaonK5hWzTXIxkiAsDIz34O4iiIdvMs5llBtQr1YOw/W05p9lBfHy/EUHMPGV6JOWJbEBEVMs5xh7W9HqnEpvIIJsRhCsUZULgbyKId4KNTQRpnaLrB45HBTVE/FNt3Ag8BiK2ViwbPUVTvdfAUS/ZGBN1KmEwRV4FAZVTuC6VHHnysFBYjPfHxWpou2tbp9NvP8UoYwe449rQ87h0yVwBjQsNCP2FM07+HxbvrA6EqoztzSp7nnMQWAc/0kHHL2qaz6s3R70FsTOKH1hM= ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-07T10:08:39.546 INFO:teuthology.orchestra.run.vm10.stdout:ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDL/lIbyqWb7EsV3rPP0RxJyou5X09+3tos3a8YCDBRsqpn7OqCw/6+pCjMY9sXvIHBoWqH3Ygqz0cTTsmi7UPsj25NSHtIfZNjMF3eJB7GtzRaicrPWmA15GAvYjH2n+kWA7YH6TqkN/AlWYifSjQqWB9EJCI+rNLmAv3RfR3g1WsdrTTIfyqWmLX8puKKMNCVwv72+ljfdaABGlJKiJWq6QCKbr4lG42OFKIB+XBIw+9U5AaonK5hWzTXIxkiAsDIz34O4iiIdvMs5llBtQr1YOw/W05p9lBfHy/EUHMPGV6JOWJbEBEVMs5xh7W9HqnEpvIIJsRhCsUZULgbyKId4KNTQRpnaLrB45HBTVE/FNt3Ag8BiK2ViwbPUVTvdfAUS/ZGBN1KmEwRV4FAZVTuC6VHHnysFBYjPfHxWpou2tbp9NvP8UoYwe449rQ87h0yVwBjQsNCP2FM07+HxbvrA6EqoztzSp7nnMQWAc/0kHHL2qaz6s3R70FsTOKH1hM= ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:08:39.550 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph config set mgr mgr/cephadm/allow_ptrace true 2026-03-07T10:08:39.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:39 vm02 bash[17011]: audit 2026-03-07T10:08:38.135591+0000 mgr.a (mgr.14156) 9 : audit [DBG] from='client.14170 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:08:39.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:39 vm02 bash[17011]: audit 2026-03-07T10:08:38.135591+0000 mgr.a (mgr.14156) 9 : audit [DBG] from='client.14170 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:08:39.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:39 vm02 bash[17011]: cluster 2026-03-07T10:08:38.747742+0000 mon.a (mon.0) 95 : cluster [DBG] mgrmap e12: a(active, since 2s) 2026-03-07T10:08:39.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:39 vm02 bash[17011]: cluster 2026-03-07T10:08:38.747742+0000 mon.a (mon.0) 95 : cluster [DBG] mgrmap e12: a(active, since 2s) 2026-03-07T10:08:39.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:39 vm02 bash[17011]: audit 2026-03-07T10:08:39.151341+0000 mon.a (mon.0) 96 : audit [INF] from='client.? 192.168.123.102:0/1976839056' entity='client.admin' 2026-03-07T10:08:39.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:39 vm02 bash[17011]: audit 2026-03-07T10:08:39.151341+0000 mon.a (mon.0) 96 : audit [INF] from='client.? 192.168.123.102:0/1976839056' entity='client.admin' 2026-03-07T10:08:42.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:42 vm02 bash[17011]: audit 2026-03-07T10:08:41.479014+0000 mon.a (mon.0) 97 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:42.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:42 vm02 bash[17011]: audit 2026-03-07T10:08:41.479014+0000 mon.a (mon.0) 97 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:42.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:42 vm02 bash[17011]: audit 2026-03-07T10:08:42.077509+0000 mon.a (mon.0) 98 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:42.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:42 vm02 bash[17011]: audit 2026-03-07T10:08:42.077509+0000 mon.a (mon.0) 98 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:44.304 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config 2026-03-07T10:08:44.358 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:44 vm02 bash[17011]: cluster 2026-03-07T10:08:43.080937+0000 mon.a (mon.0) 99 : cluster [DBG] mgrmap e13: a(active, since 6s) 2026-03-07T10:08:44.358 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:44 vm02 bash[17011]: cluster 2026-03-07T10:08:43.080937+0000 mon.a (mon.0) 99 : cluster [DBG] mgrmap e13: a(active, since 6s) 2026-03-07T10:08:44.726 INFO:tasks.cephadm:Distributing conf and client.admin keyring to all hosts + 0755 2026-03-07T10:08:44.726 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph orch client-keyring set client.admin '*' --mode 0755 2026-03-07T10:08:45.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:45 vm02 bash[17011]: audit 2026-03-07T10:08:44.657961+0000 mon.a (mon.0) 100 : audit [INF] from='client.? 192.168.123.102:0/975395568' entity='client.admin' 2026-03-07T10:08:45.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:45 vm02 bash[17011]: audit 2026-03-07T10:08:44.657961+0000 mon.a (mon.0) 100 : audit [INF] from='client.? 192.168.123.102:0/975395568' entity='client.admin' 2026-03-07T10:08:49.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:49 vm02 bash[17011]: audit 2026-03-07T10:08:48.189180+0000 mon.a (mon.0) 101 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:49.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:49 vm02 bash[17011]: audit 2026-03-07T10:08:48.189180+0000 mon.a (mon.0) 101 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:49.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:49 vm02 bash[17011]: audit 2026-03-07T10:08:48.192240+0000 mon.a (mon.0) 102 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:49.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:49 vm02 bash[17011]: audit 2026-03-07T10:08:48.192240+0000 mon.a (mon.0) 102 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:49.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:49 vm02 bash[17011]: audit 2026-03-07T10:08:48.193133+0000 mon.a (mon.0) 103 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm02", "name": "osd_memory_target"}]: dispatch 2026-03-07T10:08:49.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:49 vm02 bash[17011]: audit 2026-03-07T10:08:48.193133+0000 mon.a (mon.0) 103 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm02", "name": "osd_memory_target"}]: dispatch 2026-03-07T10:08:49.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:49 vm02 bash[17011]: audit 2026-03-07T10:08:48.196050+0000 mon.a (mon.0) 104 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:49.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:49 vm02 bash[17011]: audit 2026-03-07T10:08:48.196050+0000 mon.a (mon.0) 104 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:49.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:49 vm02 bash[17011]: audit 2026-03-07T10:08:48.201815+0000 mon.a (mon.0) 105 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:08:49.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:49 vm02 bash[17011]: audit 2026-03-07T10:08:48.201815+0000 mon.a (mon.0) 105 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:08:49.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:49 vm02 bash[17011]: audit 2026-03-07T10:08:48.205129+0000 mon.a (mon.0) 106 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:49.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:49 vm02 bash[17011]: audit 2026-03-07T10:08:48.205129+0000 mon.a (mon.0) 106 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:50.324 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config 2026-03-07T10:08:50.768 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm10 2026-03-07T10:08:50.768 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-07T10:08:50.768 DEBUG:teuthology.orchestra.run.vm10:> dd of=/etc/ceph/ceph.conf 2026-03-07T10:08:50.772 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-07T10:08:50.772 DEBUG:teuthology.orchestra.run.vm10:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-07T10:08:50.819 INFO:tasks.cephadm:Adding host vm10 to orchestrator... 2026-03-07T10:08:50.819 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph orch host add vm10 2026-03-07T10:08:51.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:51 vm02 bash[17011]: audit 2026-03-07T10:08:50.663701+0000 mgr.a (mgr.14156) 10 : audit [DBG] from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:08:51.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:51 vm02 bash[17011]: audit 2026-03-07T10:08:50.663701+0000 mgr.a (mgr.14156) 10 : audit [DBG] from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:08:51.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:51 vm02 bash[17011]: audit 2026-03-07T10:08:50.667343+0000 mon.a (mon.0) 107 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:51.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:51 vm02 bash[17011]: audit 2026-03-07T10:08:50.667343+0000 mon.a (mon.0) 107 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:51.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:51 vm02 bash[17011]: audit 2026-03-07T10:08:50.668378+0000 mon.a (mon.0) 108 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:08:51.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:51 vm02 bash[17011]: audit 2026-03-07T10:08:50.668378+0000 mon.a (mon.0) 108 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:08:51.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:51 vm02 bash[17011]: audit 2026-03-07T10:08:50.669633+0000 mon.a (mon.0) 109 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:08:51.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:51 vm02 bash[17011]: audit 2026-03-07T10:08:50.669633+0000 mon.a (mon.0) 109 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:08:51.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:51 vm02 bash[17011]: audit 2026-03-07T10:08:50.670344+0000 mon.a (mon.0) 110 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:08:51.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:51 vm02 bash[17011]: audit 2026-03-07T10:08:50.670344+0000 mon.a (mon.0) 110 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:08:51.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:51 vm02 bash[17011]: cephadm 2026-03-07T10:08:50.671279+0000 mgr.a (mgr.14156) 11 : cephadm [INF] Updating vm02:/etc/ceph/ceph.conf 2026-03-07T10:08:51.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:51 vm02 bash[17011]: cephadm 2026-03-07T10:08:50.671279+0000 mgr.a (mgr.14156) 11 : cephadm [INF] Updating vm02:/etc/ceph/ceph.conf 2026-03-07T10:08:51.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:51 vm02 bash[17011]: cephadm 2026-03-07T10:08:50.715661+0000 mgr.a (mgr.14156) 12 : cephadm [INF] Updating vm02:/var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/config/ceph.conf 2026-03-07T10:08:51.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:51 vm02 bash[17011]: cephadm 2026-03-07T10:08:50.715661+0000 mgr.a (mgr.14156) 12 : cephadm [INF] Updating vm02:/var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/config/ceph.conf 2026-03-07T10:08:51.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:51 vm02 bash[17011]: cephadm 2026-03-07T10:08:50.759978+0000 mgr.a (mgr.14156) 13 : cephadm [INF] Updating vm02:/etc/ceph/ceph.client.admin.keyring 2026-03-07T10:08:51.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:51 vm02 bash[17011]: cephadm 2026-03-07T10:08:50.759978+0000 mgr.a (mgr.14156) 13 : cephadm [INF] Updating vm02:/etc/ceph/ceph.client.admin.keyring 2026-03-07T10:08:51.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:51 vm02 bash[17011]: cephadm 2026-03-07T10:08:50.794681+0000 mgr.a (mgr.14156) 14 : cephadm [INF] Updating vm02:/var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/config/ceph.client.admin.keyring 2026-03-07T10:08:51.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:51 vm02 bash[17011]: cephadm 2026-03-07T10:08:50.794681+0000 mgr.a (mgr.14156) 14 : cephadm [INF] Updating vm02:/var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/config/ceph.client.admin.keyring 2026-03-07T10:08:51.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:51 vm02 bash[17011]: audit 2026-03-07T10:08:50.832390+0000 mon.a (mon.0) 111 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:51.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:51 vm02 bash[17011]: audit 2026-03-07T10:08:50.832390+0000 mon.a (mon.0) 111 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:51.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:51 vm02 bash[17011]: audit 2026-03-07T10:08:50.834961+0000 mon.a (mon.0) 112 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:51.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:51 vm02 bash[17011]: audit 2026-03-07T10:08:50.834961+0000 mon.a (mon.0) 112 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:51.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:51 vm02 bash[17011]: audit 2026-03-07T10:08:50.841266+0000 mon.a (mon.0) 113 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:51.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:51 vm02 bash[17011]: audit 2026-03-07T10:08:50.841266+0000 mon.a (mon.0) 113 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:55.585 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config 2026-03-07T10:08:56.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:56 vm02 bash[17011]: audit 2026-03-07T10:08:55.946978+0000 mgr.a (mgr.14156) 15 : audit [DBG] from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm10", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:08:56.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:56 vm02 bash[17011]: audit 2026-03-07T10:08:55.946978+0000 mgr.a (mgr.14156) 15 : audit [DBG] from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm10", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:08:57.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:57 vm02 bash[17011]: cluster 2026-03-07T10:08:56.268692+0000 mgr.a (mgr.14156) 16 : cluster [DBG] pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:08:57.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:57 vm02 bash[17011]: cluster 2026-03-07T10:08:56.268692+0000 mgr.a (mgr.14156) 16 : cluster [DBG] pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:08:57.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:57 vm02 bash[17011]: cephadm 2026-03-07T10:08:56.824330+0000 mgr.a (mgr.14156) 17 : cephadm [INF] Deploying cephadm binary to vm10 2026-03-07T10:08:57.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:57 vm02 bash[17011]: cephadm 2026-03-07T10:08:56.824330+0000 mgr.a (mgr.14156) 17 : cephadm [INF] Deploying cephadm binary to vm10 2026-03-07T10:08:58.612 INFO:teuthology.orchestra.run.vm02.stdout:Added host 'vm10' with addr '192.168.123.110' 2026-03-07T10:08:58.689 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph orch host ls --format=json 2026-03-07T10:08:59.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:59 vm02 bash[17011]: cluster 2026-03-07T10:08:58.268941+0000 mgr.a (mgr.14156) 18 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:08:59.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:59 vm02 bash[17011]: cluster 2026-03-07T10:08:58.268941+0000 mgr.a (mgr.14156) 18 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:08:59.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:59 vm02 bash[17011]: audit 2026-03-07T10:08:58.610958+0000 mon.a (mon.0) 114 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:59.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:59 vm02 bash[17011]: audit 2026-03-07T10:08:58.610958+0000 mon.a (mon.0) 114 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:59.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:59 vm02 bash[17011]: cephadm 2026-03-07T10:08:58.611454+0000 mgr.a (mgr.14156) 19 : cephadm [INF] Added host vm10 2026-03-07T10:08:59.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:59 vm02 bash[17011]: cephadm 2026-03-07T10:08:58.611454+0000 mgr.a (mgr.14156) 19 : cephadm [INF] Added host vm10 2026-03-07T10:08:59.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:59 vm02 bash[17011]: audit 2026-03-07T10:08:58.614356+0000 mon.a (mon.0) 115 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:08:59.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:59 vm02 bash[17011]: audit 2026-03-07T10:08:58.614356+0000 mon.a (mon.0) 115 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:08:59.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:59 vm02 bash[17011]: audit 2026-03-07T10:08:58.905377+0000 mon.a (mon.0) 116 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:08:59.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:08:59 vm02 bash[17011]: audit 2026-03-07T10:08:58.905377+0000 mon.a (mon.0) 116 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:01.461 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:01 vm02 bash[17011]: audit 2026-03-07T10:09:00.190794+0000 mon.a (mon.0) 117 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:01.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:01 vm02 bash[17011]: audit 2026-03-07T10:09:00.190794+0000 mon.a (mon.0) 117 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:01.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:01 vm02 bash[17011]: cluster 2026-03-07T10:09:00.269240+0000 mgr.a (mgr.14156) 20 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:01.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:01 vm02 bash[17011]: cluster 2026-03-07T10:09:00.269240+0000 mgr.a (mgr.14156) 20 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:01.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:01 vm02 bash[17011]: audit 2026-03-07T10:09:00.776089+0000 mon.a (mon.0) 118 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:01.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:01 vm02 bash[17011]: audit 2026-03-07T10:09:00.776089+0000 mon.a (mon.0) 118 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:03.452 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config 2026-03-07T10:09:03.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:03 vm02 bash[17011]: cluster 2026-03-07T10:09:02.269422+0000 mgr.a (mgr.14156) 21 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:03.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:03 vm02 bash[17011]: cluster 2026-03-07T10:09:02.269422+0000 mgr.a (mgr.14156) 21 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:03.846 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-07T10:09:03.846 INFO:teuthology.orchestra.run.vm02.stdout:[{"addr": "192.168.123.102", "hostname": "vm02", "labels": [], "status": ""}, {"addr": "192.168.123.110", "hostname": "vm10", "labels": [], "status": ""}] 2026-03-07T10:09:04.437 INFO:tasks.cephadm:Setting crush tunables to default 2026-03-07T10:09:04.438 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph osd crush tunables default 2026-03-07T10:09:04.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:04 vm02 bash[17011]: audit 2026-03-07T10:09:03.844757+0000 mgr.a (mgr.14156) 22 : audit [DBG] from='client.14182 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T10:09:04.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:04 vm02 bash[17011]: audit 2026-03-07T10:09:03.844757+0000 mgr.a (mgr.14156) 22 : audit [DBG] from='client.14182 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T10:09:05.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: cluster 2026-03-07T10:09:04.269579+0000 mgr.a (mgr.14156) 23 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:05.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: cluster 2026-03-07T10:09:04.269579+0000 mgr.a (mgr.14156) 23 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:05.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: audit 2026-03-07T10:09:04.491037+0000 mon.a (mon.0) 119 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:05.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: audit 2026-03-07T10:09:04.491037+0000 mon.a (mon.0) 119 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:05.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: audit 2026-03-07T10:09:04.493933+0000 mon.a (mon.0) 120 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:05.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: audit 2026-03-07T10:09:04.493933+0000 mon.a (mon.0) 120 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:05.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: audit 2026-03-07T10:09:04.497362+0000 mon.a (mon.0) 121 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:05.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: audit 2026-03-07T10:09:04.497362+0000 mon.a (mon.0) 121 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:05.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: audit 2026-03-07T10:09:04.499886+0000 mon.a (mon.0) 122 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:05.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: audit 2026-03-07T10:09:04.499886+0000 mon.a (mon.0) 122 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:05.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: audit 2026-03-07T10:09:04.500780+0000 mon.a (mon.0) 123 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-07T10:09:05.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: audit 2026-03-07T10:09:04.500780+0000 mon.a (mon.0) 123 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-07T10:09:05.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: audit 2026-03-07T10:09:04.501421+0000 mon.a (mon.0) 124 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:05.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: audit 2026-03-07T10:09:04.501421+0000 mon.a (mon.0) 124 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:05.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: audit 2026-03-07T10:09:04.502179+0000 mon.a (mon.0) 125 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:09:05.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: audit 2026-03-07T10:09:04.502179+0000 mon.a (mon.0) 125 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:09:05.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: cephadm 2026-03-07T10:09:04.503011+0000 mgr.a (mgr.14156) 24 : cephadm [INF] Updating vm10:/etc/ceph/ceph.conf 2026-03-07T10:09:05.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: cephadm 2026-03-07T10:09:04.503011+0000 mgr.a (mgr.14156) 24 : cephadm [INF] Updating vm10:/etc/ceph/ceph.conf 2026-03-07T10:09:05.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: cephadm 2026-03-07T10:09:04.534251+0000 mgr.a (mgr.14156) 25 : cephadm [INF] Updating vm10:/var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/config/ceph.conf 2026-03-07T10:09:05.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: cephadm 2026-03-07T10:09:04.534251+0000 mgr.a (mgr.14156) 25 : cephadm [INF] Updating vm10:/var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/config/ceph.conf 2026-03-07T10:09:05.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: cephadm 2026-03-07T10:09:04.565153+0000 mgr.a (mgr.14156) 26 : cephadm [INF] Updating vm10:/etc/ceph/ceph.client.admin.keyring 2026-03-07T10:09:05.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: cephadm 2026-03-07T10:09:04.565153+0000 mgr.a (mgr.14156) 26 : cephadm [INF] Updating vm10:/etc/ceph/ceph.client.admin.keyring 2026-03-07T10:09:05.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: cephadm 2026-03-07T10:09:04.596328+0000 mgr.a (mgr.14156) 27 : cephadm [INF] Updating vm10:/var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/config/ceph.client.admin.keyring 2026-03-07T10:09:05.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: cephadm 2026-03-07T10:09:04.596328+0000 mgr.a (mgr.14156) 27 : cephadm [INF] Updating vm10:/var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/config/ceph.client.admin.keyring 2026-03-07T10:09:05.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: audit 2026-03-07T10:09:04.626586+0000 mon.a (mon.0) 126 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:05.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: audit 2026-03-07T10:09:04.626586+0000 mon.a (mon.0) 126 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:05.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: audit 2026-03-07T10:09:04.634213+0000 mon.a (mon.0) 127 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:05.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: audit 2026-03-07T10:09:04.634213+0000 mon.a (mon.0) 127 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:05.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: audit 2026-03-07T10:09:04.641972+0000 mon.a (mon.0) 128 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:05.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:05 vm02 bash[17011]: audit 2026-03-07T10:09:04.641972+0000 mon.a (mon.0) 128 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:07.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:07 vm02 bash[17011]: cluster 2026-03-07T10:09:06.269810+0000 mgr.a (mgr.14156) 28 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:07.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:07 vm02 bash[17011]: cluster 2026-03-07T10:09:06.269810+0000 mgr.a (mgr.14156) 28 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:09.219 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config 2026-03-07T10:09:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:09 vm02 bash[17011]: cluster 2026-03-07T10:09:08.269984+0000 mgr.a (mgr.14156) 29 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:09 vm02 bash[17011]: cluster 2026-03-07T10:09:08.269984+0000 mgr.a (mgr.14156) 29 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:10.505 INFO:teuthology.orchestra.run.vm02.stderr:adjusted tunables profile to default 2026-03-07T10:09:10.577 INFO:tasks.cephadm:Adding mon.a on vm02 2026-03-07T10:09:10.577 INFO:tasks.cephadm:Adding mon.b on vm10 2026-03-07T10:09:10.577 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph orch apply mon '2;vm02:192.168.123.102=a;vm10:192.168.123.110=b' 2026-03-07T10:09:10.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:10 vm02 bash[17011]: audit 2026-03-07T10:09:09.563085+0000 mon.a (mon.0) 129 : audit [INF] from='client.? 192.168.123.102:0/2004074348' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-07T10:09:10.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:10 vm02 bash[17011]: audit 2026-03-07T10:09:09.563085+0000 mon.a (mon.0) 129 : audit [INF] from='client.? 192.168.123.102:0/2004074348' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-07T10:09:11.842 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/config/ceph.conf 2026-03-07T10:09:11.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:11 vm02 bash[17011]: cluster 2026-03-07T10:09:10.270226+0000 mgr.a (mgr.14156) 30 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:11.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:11 vm02 bash[17011]: cluster 2026-03-07T10:09:10.270226+0000 mgr.a (mgr.14156) 30 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:11.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:11 vm02 bash[17011]: audit 2026-03-07T10:09:10.504674+0000 mon.a (mon.0) 130 : audit [INF] from='client.? 192.168.123.102:0/2004074348' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-07T10:09:11.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:11 vm02 bash[17011]: audit 2026-03-07T10:09:10.504674+0000 mon.a (mon.0) 130 : audit [INF] from='client.? 192.168.123.102:0/2004074348' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-07T10:09:11.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:11 vm02 bash[17011]: cluster 2026-03-07T10:09:10.507140+0000 mon.a (mon.0) 131 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-07T10:09:11.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:11 vm02 bash[17011]: cluster 2026-03-07T10:09:10.507140+0000 mon.a (mon.0) 131 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-07T10:09:12.255 INFO:teuthology.orchestra.run.vm10.stdout:Scheduled mon update... 2026-03-07T10:09:12.498 DEBUG:teuthology.orchestra.run.vm10:mon.b> sudo journalctl -f -n 0 -u ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@mon.b.service 2026-03-07T10:09:12.499 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-07T10:09:12.499 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph mon dump -f json 2026-03-07T10:09:13.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:13 vm02 bash[17011]: audit 2026-03-07T10:09:12.193450+0000 mgr.a (mgr.14156) 31 : audit [DBG] from='client.14186 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "2;vm02:192.168.123.102=a;vm10:192.168.123.110=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:09:13.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:13 vm02 bash[17011]: audit 2026-03-07T10:09:12.193450+0000 mgr.a (mgr.14156) 31 : audit [DBG] from='client.14186 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "2;vm02:192.168.123.102=a;vm10:192.168.123.110=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:09:13.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:13 vm02 bash[17011]: cephadm 2026-03-07T10:09:12.194578+0000 mgr.a (mgr.14156) 32 : cephadm [INF] Saving service mon spec with placement vm02:192.168.123.102=a;vm10:192.168.123.110=b;count:2 2026-03-07T10:09:13.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:13 vm02 bash[17011]: cephadm 2026-03-07T10:09:12.194578+0000 mgr.a (mgr.14156) 32 : cephadm [INF] Saving service mon spec with placement vm02:192.168.123.102=a;vm10:192.168.123.110=b;count:2 2026-03-07T10:09:13.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:13 vm02 bash[17011]: audit 2026-03-07T10:09:12.253324+0000 mon.a (mon.0) 132 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:13.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:13 vm02 bash[17011]: audit 2026-03-07T10:09:12.253324+0000 mon.a (mon.0) 132 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:13.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:13 vm02 bash[17011]: audit 2026-03-07T10:09:12.254076+0000 mon.a (mon.0) 133 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:09:13.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:13 vm02 bash[17011]: audit 2026-03-07T10:09:12.254076+0000 mon.a (mon.0) 133 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:09:13.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:13 vm02 bash[17011]: audit 2026-03-07T10:09:12.255219+0000 mon.a (mon.0) 134 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:13.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:13 vm02 bash[17011]: audit 2026-03-07T10:09:12.255219+0000 mon.a (mon.0) 134 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:13.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:13 vm02 bash[17011]: audit 2026-03-07T10:09:12.255721+0000 mon.a (mon.0) 135 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:09:13.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:13 vm02 bash[17011]: audit 2026-03-07T10:09:12.255721+0000 mon.a (mon.0) 135 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:09:13.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:13 vm02 bash[17011]: cluster 2026-03-07T10:09:12.270487+0000 mgr.a (mgr.14156) 33 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:13.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:13 vm02 bash[17011]: cluster 2026-03-07T10:09:12.270487+0000 mgr.a (mgr.14156) 33 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:13.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:13 vm02 bash[17011]: audit 2026-03-07T10:09:12.303430+0000 mon.a (mon.0) 136 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:13.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:13 vm02 bash[17011]: audit 2026-03-07T10:09:12.303430+0000 mon.a (mon.0) 136 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:13.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:13 vm02 bash[17011]: audit 2026-03-07T10:09:12.305444+0000 mon.a (mon.0) 137 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T10:09:13.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:13 vm02 bash[17011]: audit 2026-03-07T10:09:12.305444+0000 mon.a (mon.0) 137 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T10:09:13.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:13 vm02 bash[17011]: audit 2026-03-07T10:09:12.306460+0000 mon.a (mon.0) 138 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:13.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:13 vm02 bash[17011]: audit 2026-03-07T10:09:12.306460+0000 mon.a (mon.0) 138 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:13.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:13 vm02 bash[17011]: cephadm 2026-03-07T10:09:12.307445+0000 mgr.a (mgr.14156) 34 : cephadm [INF] Deploying daemon mon.b on vm10 2026-03-07T10:09:13.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:13 vm02 bash[17011]: cephadm 2026-03-07T10:09:12.307445+0000 mgr.a (mgr.14156) 34 : cephadm [INF] Deploying daemon mon.b on vm10 2026-03-07T10:09:13.861 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.b/config 2026-03-07T10:09:14.446 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:14 vm10 bash[19283]: debug 2026-03-07T10:09:14.316+0000 7f7c2b232640 4 mon.b@-1(synchronizing).mgr e13 mkfs or daemon transitioned to available, loading commands 2026-03-07T10:09:19.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:14.270658+0000 mgr.a (mgr.14156) 35 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:19.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:14.270658+0000 mgr.a (mgr.14156) 35 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:19.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: audit 2026-03-07T10:09:14.327545+0000 mon.a (mon.0) 145 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-07T10:09:19.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: audit 2026-03-07T10:09:14.327545+0000 mon.a (mon.0) 145 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-07T10:09:19.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: audit 2026-03-07T10:09:14.327612+0000 mon.a (mon.0) 146 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:09:19.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: audit 2026-03-07T10:09:14.327612+0000 mon.a (mon.0) 146 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:09:19.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:14.327743+0000 mon.a (mon.0) 147 : cluster [INF] mon.a calling monitor election 2026-03-07T10:09:19.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:14.327743+0000 mon.a (mon.0) 147 : cluster [INF] mon.a calling monitor election 2026-03-07T10:09:19.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: audit 2026-03-07T10:09:15.323393+0000 mon.a (mon.0) 148 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:09:19.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: audit 2026-03-07T10:09:15.323393+0000 mon.a (mon.0) 148 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:09:19.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:16.270817+0000 mgr.a (mgr.14156) 36 : cluster [DBG] pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:19.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:16.270817+0000 mgr.a (mgr.14156) 36 : cluster [DBG] pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:19.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: audit 2026-03-07T10:09:16.323303+0000 mon.a (mon.0) 149 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:09:19.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: audit 2026-03-07T10:09:16.323303+0000 mon.a (mon.0) 149 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:09:19.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:16.326451+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-07T10:09:19.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:16.326451+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-07T10:09:19.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: audit 2026-03-07T10:09:17.323413+0000 mon.a (mon.0) 150 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:09:19.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: audit 2026-03-07T10:09:17.323413+0000 mon.a (mon.0) 150 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:09:19.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:18.271007+0000 mgr.a (mgr.14156) 37 : cluster [DBG] pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:19.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:18.271007+0000 mgr.a (mgr.14156) 37 : cluster [DBG] pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: audit 2026-03-07T10:09:18.323526+0000 mon.a (mon.0) 151 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: audit 2026-03-07T10:09:18.323526+0000 mon.a (mon.0) 151 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: audit 2026-03-07T10:09:19.323644+0000 mon.a (mon.0) 152 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: audit 2026-03-07T10:09:19.323644+0000 mon.a (mon.0) 152 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:19.333086+0000 mon.a (mon.0) 153 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:19.333086+0000 mon.a (mon.0) 153 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:19.336250+0000 mon.a (mon.0) 154 : cluster [DBG] monmap epoch 2 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:19.336250+0000 mon.a (mon.0) 154 : cluster [DBG] monmap epoch 2 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:19.336285+0000 mon.a (mon.0) 155 : cluster [DBG] fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:19.336285+0000 mon.a (mon.0) 155 : cluster [DBG] fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:19.336298+0000 mon.a (mon.0) 156 : cluster [DBG] last_changed 2026-03-07T10:09:14.324495+0000 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:19.336298+0000 mon.a (mon.0) 156 : cluster [DBG] last_changed 2026-03-07T10:09:14.324495+0000 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:19.336310+0000 mon.a (mon.0) 157 : cluster [DBG] created 2026-03-07T10:07:51.937005+0000 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:19.336310+0000 mon.a (mon.0) 157 : cluster [DBG] created 2026-03-07T10:07:51.937005+0000 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:19.336319+0000 mon.a (mon.0) 158 : cluster [DBG] min_mon_release 19 (squid) 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:19.336319+0000 mon.a (mon.0) 158 : cluster [DBG] min_mon_release 19 (squid) 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:19.336331+0000 mon.a (mon.0) 159 : cluster [DBG] election_strategy: 1 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:19.336331+0000 mon.a (mon.0) 159 : cluster [DBG] election_strategy: 1 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:19.336345+0000 mon.a (mon.0) 160 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:19.336345+0000 mon.a (mon.0) 160 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:19.336361+0000 mon.a (mon.0) 161 : cluster [DBG] 1: [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon.b 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:19.336361+0000 mon.a (mon.0) 161 : cluster [DBG] 1: [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon.b 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:19.336731+0000 mon.a (mon.0) 162 : cluster [DBG] fsmap 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:19.336731+0000 mon.a (mon.0) 162 : cluster [DBG] fsmap 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:19.336754+0000 mon.a (mon.0) 163 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:19.336754+0000 mon.a (mon.0) 163 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:19.336887+0000 mon.a (mon.0) 164 : cluster [DBG] mgrmap e13: a(active, since 43s) 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:19.336887+0000 mon.a (mon.0) 164 : cluster [DBG] mgrmap e13: a(active, since 43s) 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:19.336979+0000 mon.a (mon.0) 165 : cluster [INF] overall HEALTH_OK 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: cluster 2026-03-07T10:09:19.336979+0000 mon.a (mon.0) 165 : cluster [INF] overall HEALTH_OK 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: audit 2026-03-07T10:09:19.340729+0000 mon.a (mon.0) 166 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: audit 2026-03-07T10:09:19.340729+0000 mon.a (mon.0) 166 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: audit 2026-03-07T10:09:19.345496+0000 mon.a (mon.0) 167 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: audit 2026-03-07T10:09:19.345496+0000 mon.a (mon.0) 167 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: audit 2026-03-07T10:09:19.350478+0000 mon.a (mon.0) 168 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: audit 2026-03-07T10:09:19.350478+0000 mon.a (mon.0) 168 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:19.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: audit 2026-03-07T10:09:19.351590+0000 mon.a (mon.0) 169 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:19.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: audit 2026-03-07T10:09:19.351590+0000 mon.a (mon.0) 169 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:19.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: audit 2026-03-07T10:09:19.352212+0000 mon.a (mon.0) 170 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:09:19.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:19 vm10 bash[19283]: audit 2026-03-07T10:09:19.352212+0000 mon.a (mon.0) 170 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:09:19.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:14.270658+0000 mgr.a (mgr.14156) 35 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:19.714 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:14.270658+0000 mgr.a (mgr.14156) 35 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:19.714 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: audit 2026-03-07T10:09:14.327545+0000 mon.a (mon.0) 145 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-07T10:09:19.714 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: audit 2026-03-07T10:09:14.327545+0000 mon.a (mon.0) 145 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-07T10:09:19.714 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: audit 2026-03-07T10:09:14.327612+0000 mon.a (mon.0) 146 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:09:19.714 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: audit 2026-03-07T10:09:14.327612+0000 mon.a (mon.0) 146 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:09:19.714 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:14.327743+0000 mon.a (mon.0) 147 : cluster [INF] mon.a calling monitor election 2026-03-07T10:09:19.714 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:14.327743+0000 mon.a (mon.0) 147 : cluster [INF] mon.a calling monitor election 2026-03-07T10:09:19.714 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: audit 2026-03-07T10:09:15.323393+0000 mon.a (mon.0) 148 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:09:19.714 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: audit 2026-03-07T10:09:15.323393+0000 mon.a (mon.0) 148 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:09:19.714 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:16.270817+0000 mgr.a (mgr.14156) 36 : cluster [DBG] pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:19.714 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:16.270817+0000 mgr.a (mgr.14156) 36 : cluster [DBG] pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:19.714 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: audit 2026-03-07T10:09:16.323303+0000 mon.a (mon.0) 149 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:09:19.714 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: audit 2026-03-07T10:09:16.323303+0000 mon.a (mon.0) 149 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:09:19.714 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:16.326451+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-07T10:09:19.714 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:16.326451+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-07T10:09:19.714 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: audit 2026-03-07T10:09:17.323413+0000 mon.a (mon.0) 150 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: audit 2026-03-07T10:09:17.323413+0000 mon.a (mon.0) 150 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:18.271007+0000 mgr.a (mgr.14156) 37 : cluster [DBG] pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:18.271007+0000 mgr.a (mgr.14156) 37 : cluster [DBG] pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: audit 2026-03-07T10:09:18.323526+0000 mon.a (mon.0) 151 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: audit 2026-03-07T10:09:18.323526+0000 mon.a (mon.0) 151 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: audit 2026-03-07T10:09:19.323644+0000 mon.a (mon.0) 152 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: audit 2026-03-07T10:09:19.323644+0000 mon.a (mon.0) 152 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:19.333086+0000 mon.a (mon.0) 153 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:19.333086+0000 mon.a (mon.0) 153 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:19.336250+0000 mon.a (mon.0) 154 : cluster [DBG] monmap epoch 2 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:19.336250+0000 mon.a (mon.0) 154 : cluster [DBG] monmap epoch 2 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:19.336285+0000 mon.a (mon.0) 155 : cluster [DBG] fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:19.336285+0000 mon.a (mon.0) 155 : cluster [DBG] fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:19.336298+0000 mon.a (mon.0) 156 : cluster [DBG] last_changed 2026-03-07T10:09:14.324495+0000 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:19.336298+0000 mon.a (mon.0) 156 : cluster [DBG] last_changed 2026-03-07T10:09:14.324495+0000 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:19.336310+0000 mon.a (mon.0) 157 : cluster [DBG] created 2026-03-07T10:07:51.937005+0000 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:19.336310+0000 mon.a (mon.0) 157 : cluster [DBG] created 2026-03-07T10:07:51.937005+0000 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:19.336319+0000 mon.a (mon.0) 158 : cluster [DBG] min_mon_release 19 (squid) 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:19.336319+0000 mon.a (mon.0) 158 : cluster [DBG] min_mon_release 19 (squid) 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:19.336331+0000 mon.a (mon.0) 159 : cluster [DBG] election_strategy: 1 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:19.336331+0000 mon.a (mon.0) 159 : cluster [DBG] election_strategy: 1 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:19.336345+0000 mon.a (mon.0) 160 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:19.336345+0000 mon.a (mon.0) 160 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:19.336361+0000 mon.a (mon.0) 161 : cluster [DBG] 1: [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon.b 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:19.336361+0000 mon.a (mon.0) 161 : cluster [DBG] 1: [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon.b 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:19.336731+0000 mon.a (mon.0) 162 : cluster [DBG] fsmap 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:19.336731+0000 mon.a (mon.0) 162 : cluster [DBG] fsmap 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:19.336754+0000 mon.a (mon.0) 163 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:19.336754+0000 mon.a (mon.0) 163 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:19.336887+0000 mon.a (mon.0) 164 : cluster [DBG] mgrmap e13: a(active, since 43s) 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:19.336887+0000 mon.a (mon.0) 164 : cluster [DBG] mgrmap e13: a(active, since 43s) 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:19.336979+0000 mon.a (mon.0) 165 : cluster [INF] overall HEALTH_OK 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: cluster 2026-03-07T10:09:19.336979+0000 mon.a (mon.0) 165 : cluster [INF] overall HEALTH_OK 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: audit 2026-03-07T10:09:19.340729+0000 mon.a (mon.0) 166 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: audit 2026-03-07T10:09:19.340729+0000 mon.a (mon.0) 166 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: audit 2026-03-07T10:09:19.345496+0000 mon.a (mon.0) 167 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: audit 2026-03-07T10:09:19.345496+0000 mon.a (mon.0) 167 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: audit 2026-03-07T10:09:19.350478+0000 mon.a (mon.0) 168 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: audit 2026-03-07T10:09:19.350478+0000 mon.a (mon.0) 168 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:19.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: audit 2026-03-07T10:09:19.351590+0000 mon.a (mon.0) 169 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:19.716 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: audit 2026-03-07T10:09:19.351590+0000 mon.a (mon.0) 169 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:19.716 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: audit 2026-03-07T10:09:19.352212+0000 mon.a (mon.0) 170 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:09:19.716 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:19 vm02 bash[17011]: audit 2026-03-07T10:09:19.352212+0000 mon.a (mon.0) 170 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:09:20.475 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-07T10:09:20.475 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":2,"fsid":"6c715b7a-1a0d-11f1-b180-89615ccd948e","modified":"2026-03-07T10:09:14.324495Z","created":"2026-03-07T10:07:51.937005Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:3300","nonce":0},{"type":"v1","addr":"192.168.123.102:6789","nonce":0}]},"addr":"192.168.123.102:6789/0","public_addr":"192.168.123.102:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"b","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:3300","nonce":0},{"type":"v1","addr":"192.168.123.110:6789","nonce":0}]},"addr":"192.168.123.110:6789/0","public_addr":"192.168.123.110:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1]} 2026-03-07T10:09:20.476 INFO:teuthology.orchestra.run.vm10.stderr:dumped monmap epoch 2 2026-03-07T10:09:20.565 INFO:tasks.cephadm:Generating final ceph.conf file... 2026-03-07T10:09:20.566 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph config generate-minimal-conf 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: cephadm 2026-03-07T10:09:19.352793+0000 mgr.a (mgr.14156) 38 : cephadm [INF] Updating vm02:/etc/ceph/ceph.conf 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: cephadm 2026-03-07T10:09:19.352793+0000 mgr.a (mgr.14156) 38 : cephadm [INF] Updating vm02:/etc/ceph/ceph.conf 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: cephadm 2026-03-07T10:09:19.352892+0000 mgr.a (mgr.14156) 39 : cephadm [INF] Updating vm10:/etc/ceph/ceph.conf 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: cephadm 2026-03-07T10:09:19.352892+0000 mgr.a (mgr.14156) 39 : cephadm [INF] Updating vm10:/etc/ceph/ceph.conf 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: cephadm 2026-03-07T10:09:19.390043+0000 mgr.a (mgr.14156) 40 : cephadm [INF] Updating vm02:/var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/config/ceph.conf 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: cephadm 2026-03-07T10:09:19.390043+0000 mgr.a (mgr.14156) 40 : cephadm [INF] Updating vm02:/var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/config/ceph.conf 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: cephadm 2026-03-07T10:09:19.401933+0000 mgr.a (mgr.14156) 41 : cephadm [INF] Updating vm10:/var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/config/ceph.conf 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: cephadm 2026-03-07T10:09:19.401933+0000 mgr.a (mgr.14156) 41 : cephadm [INF] Updating vm10:/var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/config/ceph.conf 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:19.594220+0000 mon.a (mon.0) 171 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:19.594220+0000 mon.a (mon.0) 171 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:19.616722+0000 mon.a (mon.0) 172 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:19.616722+0000 mon.a (mon.0) 172 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:19.624971+0000 mon.a (mon.0) 173 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:19.624971+0000 mon.a (mon.0) 173 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:19.642463+0000 mon.a (mon.0) 174 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:19.642463+0000 mon.a (mon.0) 174 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:19.669540+0000 mon.a (mon.0) 175 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:19.669540+0000 mon.a (mon.0) 175 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:19.693558+0000 mon.a (mon.0) 176 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:19.693558+0000 mon.a (mon.0) 176 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:19.717535+0000 mon.a (mon.0) 177 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:19.717535+0000 mon.a (mon.0) 177 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:19.732325+0000 mon.a (mon.0) 178 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:19.732325+0000 mon.a (mon.0) 178 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:19.749075+0000 mon.a (mon.0) 179 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:19.749075+0000 mon.a (mon.0) 179 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: cephadm 2026-03-07T10:09:19.749662+0000 mgr.a (mgr.14156) 42 : cephadm [INF] Reconfiguring mon.a (unknown last config time)... 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: cephadm 2026-03-07T10:09:19.749662+0000 mgr.a (mgr.14156) 42 : cephadm [INF] Reconfiguring mon.a (unknown last config time)... 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:19.750023+0000 mon.a (mon.0) 180 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:19.750023+0000 mon.a (mon.0) 180 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:19.750698+0000 mon.a (mon.0) 181 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:19.750698+0000 mon.a (mon.0) 181 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:19.751213+0000 mon.a (mon.0) 182 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:19.751213+0000 mon.a (mon.0) 182 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: cephadm 2026-03-07T10:09:19.751916+0000 mgr.a (mgr.14156) 43 : cephadm [INF] Reconfiguring daemon mon.a on vm02 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: cephadm 2026-03-07T10:09:19.751916+0000 mgr.a (mgr.14156) 43 : cephadm [INF] Reconfiguring daemon mon.a on vm02 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:20.324219+0000 mon.a (mon.0) 183 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:20.324219+0000 mon.a (mon.0) 183 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:20.330989+0000 mon.a (mon.0) 184 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:20.330989+0000 mon.a (mon.0) 184 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:20.335412+0000 mon.a (mon.0) 185 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:20.335412+0000 mon.a (mon.0) 185 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:20.336304+0000 mon.a (mon.0) 186 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:20.336304+0000 mon.a (mon.0) 186 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T10:09:20.753 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:20.336753+0000 mon.a (mon.0) 187 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-07T10:09:20.754 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:20.336753+0000 mon.a (mon.0) 187 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-07T10:09:20.754 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:20.337199+0000 mon.a (mon.0) 188 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:20.754 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:20.337199+0000 mon.a (mon.0) 188 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:20.754 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:20.474519+0000 mon.a (mon.0) 189 : audit [DBG] from='client.? 192.168.123.110:0/2182041725' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-07T10:09:20.754 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:20 vm10 bash[19283]: audit 2026-03-07T10:09:20.474519+0000 mon.a (mon.0) 189 : audit [DBG] from='client.? 192.168.123.110:0/2182041725' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-07T10:09:20.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: cephadm 2026-03-07T10:09:19.352793+0000 mgr.a (mgr.14156) 38 : cephadm [INF] Updating vm02:/etc/ceph/ceph.conf 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: cephadm 2026-03-07T10:09:19.352793+0000 mgr.a (mgr.14156) 38 : cephadm [INF] Updating vm02:/etc/ceph/ceph.conf 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: cephadm 2026-03-07T10:09:19.352892+0000 mgr.a (mgr.14156) 39 : cephadm [INF] Updating vm10:/etc/ceph/ceph.conf 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: cephadm 2026-03-07T10:09:19.352892+0000 mgr.a (mgr.14156) 39 : cephadm [INF] Updating vm10:/etc/ceph/ceph.conf 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: cephadm 2026-03-07T10:09:19.390043+0000 mgr.a (mgr.14156) 40 : cephadm [INF] Updating vm02:/var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/config/ceph.conf 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: cephadm 2026-03-07T10:09:19.390043+0000 mgr.a (mgr.14156) 40 : cephadm [INF] Updating vm02:/var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/config/ceph.conf 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: cephadm 2026-03-07T10:09:19.401933+0000 mgr.a (mgr.14156) 41 : cephadm [INF] Updating vm10:/var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/config/ceph.conf 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: cephadm 2026-03-07T10:09:19.401933+0000 mgr.a (mgr.14156) 41 : cephadm [INF] Updating vm10:/var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/config/ceph.conf 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:19.594220+0000 mon.a (mon.0) 171 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:19.594220+0000 mon.a (mon.0) 171 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:19.616722+0000 mon.a (mon.0) 172 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:19.616722+0000 mon.a (mon.0) 172 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:19.624971+0000 mon.a (mon.0) 173 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:19.624971+0000 mon.a (mon.0) 173 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:19.642463+0000 mon.a (mon.0) 174 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:19.642463+0000 mon.a (mon.0) 174 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:19.669540+0000 mon.a (mon.0) 175 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:19.669540+0000 mon.a (mon.0) 175 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:19.693558+0000 mon.a (mon.0) 176 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:19.693558+0000 mon.a (mon.0) 176 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:19.717535+0000 mon.a (mon.0) 177 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:19.717535+0000 mon.a (mon.0) 177 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:19.732325+0000 mon.a (mon.0) 178 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:19.732325+0000 mon.a (mon.0) 178 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:19.749075+0000 mon.a (mon.0) 179 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:19.749075+0000 mon.a (mon.0) 179 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: cephadm 2026-03-07T10:09:19.749662+0000 mgr.a (mgr.14156) 42 : cephadm [INF] Reconfiguring mon.a (unknown last config time)... 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: cephadm 2026-03-07T10:09:19.749662+0000 mgr.a (mgr.14156) 42 : cephadm [INF] Reconfiguring mon.a (unknown last config time)... 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:19.750023+0000 mon.a (mon.0) 180 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:19.750023+0000 mon.a (mon.0) 180 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:19.750698+0000 mon.a (mon.0) 181 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:19.750698+0000 mon.a (mon.0) 181 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:19.751213+0000 mon.a (mon.0) 182 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:19.751213+0000 mon.a (mon.0) 182 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: cephadm 2026-03-07T10:09:19.751916+0000 mgr.a (mgr.14156) 43 : cephadm [INF] Reconfiguring daemon mon.a on vm02 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: cephadm 2026-03-07T10:09:19.751916+0000 mgr.a (mgr.14156) 43 : cephadm [INF] Reconfiguring daemon mon.a on vm02 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:20.324219+0000 mon.a (mon.0) 183 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:20.324219+0000 mon.a (mon.0) 183 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:20.330989+0000 mon.a (mon.0) 184 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:20.330989+0000 mon.a (mon.0) 184 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:20.335412+0000 mon.a (mon.0) 185 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:20.335412+0000 mon.a (mon.0) 185 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:20.336304+0000 mon.a (mon.0) 186 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:20.336304+0000 mon.a (mon.0) 186 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:20.336753+0000 mon.a (mon.0) 187 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:20.336753+0000 mon.a (mon.0) 187 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:20.337199+0000 mon.a (mon.0) 188 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:20.337199+0000 mon.a (mon.0) 188 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:20.474519+0000 mon.a (mon.0) 189 : audit [DBG] from='client.? 192.168.123.110:0/2182041725' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-07T10:09:20.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:20 vm02 bash[17011]: audit 2026-03-07T10:09:20.474519+0000 mon.a (mon.0) 189 : audit [DBG] from='client.? 192.168.123.110:0/2182041725' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-07T10:09:21.326 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:09:21 vm02 bash[17284]: debug 2026-03-07T10:09:21.321+0000 7f9a8867d640 -1 mgr.server handle_report got status from non-daemon mon.b 2026-03-07T10:09:22.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:21 vm10 bash[19283]: cluster 2026-03-07T10:09:20.271215+0000 mgr.a (mgr.14156) 44 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:22.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:21 vm10 bash[19283]: cluster 2026-03-07T10:09:20.271215+0000 mgr.a (mgr.14156) 44 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:22.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:21 vm10 bash[19283]: cephadm 2026-03-07T10:09:20.336118+0000 mgr.a (mgr.14156) 45 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-07T10:09:22.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:21 vm10 bash[19283]: cephadm 2026-03-07T10:09:20.336118+0000 mgr.a (mgr.14156) 45 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-07T10:09:22.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:21 vm10 bash[19283]: cephadm 2026-03-07T10:09:20.337791+0000 mgr.a (mgr.14156) 46 : cephadm [INF] Reconfiguring daemon mon.b on vm10 2026-03-07T10:09:22.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:21 vm10 bash[19283]: cephadm 2026-03-07T10:09:20.337791+0000 mgr.a (mgr.14156) 46 : cephadm [INF] Reconfiguring daemon mon.b on vm10 2026-03-07T10:09:22.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:21 vm10 bash[19283]: audit 2026-03-07T10:09:20.801305+0000 mon.a (mon.0) 190 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:22.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:21 vm10 bash[19283]: audit 2026-03-07T10:09:20.801305+0000 mon.a (mon.0) 190 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:22.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:21 vm10 bash[19283]: audit 2026-03-07T10:09:20.806082+0000 mon.a (mon.0) 191 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:22.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:21 vm10 bash[19283]: audit 2026-03-07T10:09:20.806082+0000 mon.a (mon.0) 191 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:22.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:21 vm10 bash[19283]: audit 2026-03-07T10:09:20.808175+0000 mon.a (mon.0) 192 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:09:22.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:21 vm10 bash[19283]: audit 2026-03-07T10:09:20.808175+0000 mon.a (mon.0) 192 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:09:22.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:21 vm10 bash[19283]: audit 2026-03-07T10:09:20.809197+0000 mon.a (mon.0) 193 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:22.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:21 vm10 bash[19283]: audit 2026-03-07T10:09:20.809197+0000 mon.a (mon.0) 193 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:22.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:21 vm10 bash[19283]: audit 2026-03-07T10:09:20.809610+0000 mon.a (mon.0) 194 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:09:22.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:21 vm10 bash[19283]: audit 2026-03-07T10:09:20.809610+0000 mon.a (mon.0) 194 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:09:22.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:21 vm10 bash[19283]: audit 2026-03-07T10:09:20.813899+0000 mon.a (mon.0) 195 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:22.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:21 vm10 bash[19283]: audit 2026-03-07T10:09:20.813899+0000 mon.a (mon.0) 195 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:22.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:21 vm02 bash[17011]: cluster 2026-03-07T10:09:20.271215+0000 mgr.a (mgr.14156) 44 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:22.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:21 vm02 bash[17011]: cluster 2026-03-07T10:09:20.271215+0000 mgr.a (mgr.14156) 44 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:22.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:21 vm02 bash[17011]: cephadm 2026-03-07T10:09:20.336118+0000 mgr.a (mgr.14156) 45 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-07T10:09:22.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:21 vm02 bash[17011]: cephadm 2026-03-07T10:09:20.336118+0000 mgr.a (mgr.14156) 45 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-07T10:09:22.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:21 vm02 bash[17011]: cephadm 2026-03-07T10:09:20.337791+0000 mgr.a (mgr.14156) 46 : cephadm [INF] Reconfiguring daemon mon.b on vm10 2026-03-07T10:09:22.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:21 vm02 bash[17011]: cephadm 2026-03-07T10:09:20.337791+0000 mgr.a (mgr.14156) 46 : cephadm [INF] Reconfiguring daemon mon.b on vm10 2026-03-07T10:09:22.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:21 vm02 bash[17011]: audit 2026-03-07T10:09:20.801305+0000 mon.a (mon.0) 190 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:22.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:21 vm02 bash[17011]: audit 2026-03-07T10:09:20.801305+0000 mon.a (mon.0) 190 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:22.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:21 vm02 bash[17011]: audit 2026-03-07T10:09:20.806082+0000 mon.a (mon.0) 191 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:22.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:21 vm02 bash[17011]: audit 2026-03-07T10:09:20.806082+0000 mon.a (mon.0) 191 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:22.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:21 vm02 bash[17011]: audit 2026-03-07T10:09:20.808175+0000 mon.a (mon.0) 192 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:09:22.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:21 vm02 bash[17011]: audit 2026-03-07T10:09:20.808175+0000 mon.a (mon.0) 192 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:09:22.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:21 vm02 bash[17011]: audit 2026-03-07T10:09:20.809197+0000 mon.a (mon.0) 193 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:22.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:21 vm02 bash[17011]: audit 2026-03-07T10:09:20.809197+0000 mon.a (mon.0) 193 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:22.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:21 vm02 bash[17011]: audit 2026-03-07T10:09:20.809610+0000 mon.a (mon.0) 194 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:09:22.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:21 vm02 bash[17011]: audit 2026-03-07T10:09:20.809610+0000 mon.a (mon.0) 194 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:09:22.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:21 vm02 bash[17011]: audit 2026-03-07T10:09:20.813899+0000 mon.a (mon.0) 195 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:22.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:21 vm02 bash[17011]: audit 2026-03-07T10:09:20.813899+0000 mon.a (mon.0) 195 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:24.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:23 vm10 bash[19283]: cluster 2026-03-07T10:09:22.271439+0000 mgr.a (mgr.14156) 47 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:24.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:23 vm10 bash[19283]: cluster 2026-03-07T10:09:22.271439+0000 mgr.a (mgr.14156) 47 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:24.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:23 vm02 bash[17011]: cluster 2026-03-07T10:09:22.271439+0000 mgr.a (mgr.14156) 47 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:24.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:23 vm02 bash[17011]: cluster 2026-03-07T10:09:22.271439+0000 mgr.a (mgr.14156) 47 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:25.369 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config 2026-03-07T10:09:25.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:25 vm10 bash[19283]: cluster 2026-03-07T10:09:24.271604+0000 mgr.a (mgr.14156) 48 : cluster [DBG] pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:25.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:25 vm10 bash[19283]: cluster 2026-03-07T10:09:24.271604+0000 mgr.a (mgr.14156) 48 : cluster [DBG] pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:25.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:25 vm02 bash[17011]: cluster 2026-03-07T10:09:24.271604+0000 mgr.a (mgr.14156) 48 : cluster [DBG] pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:25.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:25 vm02 bash[17011]: cluster 2026-03-07T10:09:24.271604+0000 mgr.a (mgr.14156) 48 : cluster [DBG] pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:25.741 INFO:teuthology.orchestra.run.vm02.stdout:# minimal ceph.conf for 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:09:25.741 INFO:teuthology.orchestra.run.vm02.stdout:[global] 2026-03-07T10:09:25.741 INFO:teuthology.orchestra.run.vm02.stdout: fsid = 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:09:25.741 INFO:teuthology.orchestra.run.vm02.stdout: mon_host = [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] 2026-03-07T10:09:25.810 INFO:tasks.cephadm:Distributing (final) config and client.admin keyring... 2026-03-07T10:09:25.811 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-07T10:09:25.811 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/ceph/ceph.conf 2026-03-07T10:09:25.818 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-07T10:09:25.818 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-07T10:09:25.869 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-07T10:09:25.869 DEBUG:teuthology.orchestra.run.vm10:> sudo dd of=/etc/ceph/ceph.conf 2026-03-07T10:09:25.877 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-07T10:09:25.877 DEBUG:teuthology.orchestra.run.vm10:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-07T10:09:25.927 INFO:tasks.cephadm:Adding mgr.a on vm02 2026-03-07T10:09:25.927 INFO:tasks.cephadm:Adding mgr.b on vm10 2026-03-07T10:09:25.928 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph orch apply mgr '2;vm02=a;vm10=b' 2026-03-07T10:09:26.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:26 vm10 bash[19283]: audit 2026-03-07T10:09:25.740701+0000 mon.a (mon.0) 196 : audit [DBG] from='client.? 192.168.123.102:0/4007773493' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:26.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:26 vm10 bash[19283]: audit 2026-03-07T10:09:25.740701+0000 mon.a (mon.0) 196 : audit [DBG] from='client.? 192.168.123.102:0/4007773493' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:26.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:26 vm02 bash[17011]: audit 2026-03-07T10:09:25.740701+0000 mon.a (mon.0) 196 : audit [DBG] from='client.? 192.168.123.102:0/4007773493' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:26.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:26 vm02 bash[17011]: audit 2026-03-07T10:09:25.740701+0000 mon.a (mon.0) 196 : audit [DBG] from='client.? 192.168.123.102:0/4007773493' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:27.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:27 vm10 bash[19283]: cluster 2026-03-07T10:09:26.271807+0000 mgr.a (mgr.14156) 49 : cluster [DBG] pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:27.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:27 vm10 bash[19283]: cluster 2026-03-07T10:09:26.271807+0000 mgr.a (mgr.14156) 49 : cluster [DBG] pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:27.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:27 vm02 bash[17011]: cluster 2026-03-07T10:09:26.271807+0000 mgr.a (mgr.14156) 49 : cluster [DBG] pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:27.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:27 vm02 bash[17011]: cluster 2026-03-07T10:09:26.271807+0000 mgr.a (mgr.14156) 49 : cluster [DBG] pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:29.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:29 vm10 bash[19283]: cluster 2026-03-07T10:09:28.272031+0000 mgr.a (mgr.14156) 50 : cluster [DBG] pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:29.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:29 vm10 bash[19283]: cluster 2026-03-07T10:09:28.272031+0000 mgr.a (mgr.14156) 50 : cluster [DBG] pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:29.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:29 vm02 bash[17011]: cluster 2026-03-07T10:09:28.272031+0000 mgr.a (mgr.14156) 50 : cluster [DBG] pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:29.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:29 vm02 bash[17011]: cluster 2026-03-07T10:09:28.272031+0000 mgr.a (mgr.14156) 50 : cluster [DBG] pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:29.726 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.b/config 2026-03-07T10:09:30.090 INFO:teuthology.orchestra.run.vm10.stdout:Scheduled mgr update... 2026-03-07T10:09:30.167 DEBUG:teuthology.orchestra.run.vm10:mgr.b> sudo journalctl -f -n 0 -u ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@mgr.b.service 2026-03-07T10:09:30.168 INFO:tasks.cephadm:Deploying OSDs... 2026-03-07T10:09:30.168 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-07T10:09:30.168 DEBUG:teuthology.orchestra.run.vm02:> dd if=/scratch_devs of=/dev/stdout 2026-03-07T10:09:30.170 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-07T10:09:30.171 DEBUG:teuthology.orchestra.run.vm02:> ls /dev/[sv]d? 2026-03-07T10:09:30.214 INFO:teuthology.orchestra.run.vm02.stdout:/dev/vda 2026-03-07T10:09:30.214 INFO:teuthology.orchestra.run.vm02.stdout:/dev/vdb 2026-03-07T10:09:30.214 INFO:teuthology.orchestra.run.vm02.stdout:/dev/vdc 2026-03-07T10:09:30.214 INFO:teuthology.orchestra.run.vm02.stdout:/dev/vdd 2026-03-07T10:09:30.214 INFO:teuthology.orchestra.run.vm02.stdout:/dev/vde 2026-03-07T10:09:30.214 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-07T10:09:30.214 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-07T10:09:30.214 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vdb 2026-03-07T10:09:30.258 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vdb 2026-03-07T10:09:30.259 INFO:teuthology.orchestra.run.vm02.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-07T10:09:30.259 INFO:teuthology.orchestra.run.vm02.stdout:Device: 5h/5d Inode: 24 Links: 1 Device type: fe,10 2026-03-07T10:09:30.259 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-07T10:09:30.259 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-03-07 10:03:40.051985361 +0000 2026-03-07T10:09:30.259 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-03-07 10:03:39.155985361 +0000 2026-03-07T10:09:30.259 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-03-07 10:03:39.155985361 +0000 2026-03-07T10:09:30.259 INFO:teuthology.orchestra.run.vm02.stdout: Birth: - 2026-03-07T10:09:30.259 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-07T10:09:30.305 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-03-07T10:09:30.305 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-03-07T10:09:30.305 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.000126637 s, 4.0 MB/s 2026-03-07T10:09:30.306 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-07T10:09:30.351 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vdc 2026-03-07T10:09:30.365 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:30 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:09:30.395 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vdc 2026-03-07T10:09:30.395 INFO:teuthology.orchestra.run.vm02.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-07T10:09:30.395 INFO:teuthology.orchestra.run.vm02.stdout:Device: 5h/5d Inode: 25 Links: 1 Device type: fe,20 2026-03-07T10:09:30.395 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-07T10:09:30.395 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-03-07 10:03:40.059985361 +0000 2026-03-07T10:09:30.395 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-03-07 10:03:39.159985361 +0000 2026-03-07T10:09:30.395 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-03-07 10:03:39.159985361 +0000 2026-03-07T10:09:30.395 INFO:teuthology.orchestra.run.vm02.stdout: Birth: - 2026-03-07T10:09:30.395 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-07T10:09:30.441 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-03-07T10:09:30.441 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-03-07T10:09:30.441 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.000136204 s, 3.8 MB/s 2026-03-07T10:09:30.442 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-07T10:09:30.487 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vdd 2026-03-07T10:09:30.531 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vdd 2026-03-07T10:09:30.531 INFO:teuthology.orchestra.run.vm02.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-07T10:09:30.531 INFO:teuthology.orchestra.run.vm02.stdout:Device: 5h/5d Inode: 26 Links: 1 Device type: fe,30 2026-03-07T10:09:30.531 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-07T10:09:30.531 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-03-07 10:03:40.051985361 +0000 2026-03-07T10:09:30.531 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-03-07 10:03:39.155985361 +0000 2026-03-07T10:09:30.531 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-03-07 10:03:39.155985361 +0000 2026-03-07T10:09:30.531 INFO:teuthology.orchestra.run.vm02.stdout: Birth: - 2026-03-07T10:09:30.531 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-07T10:09:30.578 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-03-07T10:09:30.578 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-03-07T10:09:30.578 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.000158025 s, 3.2 MB/s 2026-03-07T10:09:30.578 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-07T10:09:30.623 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vde 2026-03-07T10:09:30.667 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vde 2026-03-07T10:09:30.667 INFO:teuthology.orchestra.run.vm02.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-07T10:09:30.667 INFO:teuthology.orchestra.run.vm02.stdout:Device: 5h/5d Inode: 27 Links: 1 Device type: fe,40 2026-03-07T10:09:30.667 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-07T10:09:30.667 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-03-07 10:03:40.055985361 +0000 2026-03-07T10:09:30.667 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-03-07 10:03:39.159985361 +0000 2026-03-07T10:09:30.667 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-03-07 10:03:39.159985361 +0000 2026-03-07T10:09:30.667 INFO:teuthology.orchestra.run.vm02.stdout: Birth: - 2026-03-07T10:09:30.667 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-07T10:09:30.714 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-03-07T10:09:30.714 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-03-07T10:09:30.714 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.000182031 s, 2.8 MB/s 2026-03-07T10:09:30.715 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-07T10:09:30.760 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-07T10:09:30.760 DEBUG:teuthology.orchestra.run.vm10:> dd if=/scratch_devs of=/dev/stdout 2026-03-07T10:09:30.763 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-07T10:09:30.763 DEBUG:teuthology.orchestra.run.vm10:> ls /dev/[sv]d? 2026-03-07T10:09:30.811 INFO:teuthology.orchestra.run.vm10.stdout:/dev/vda 2026-03-07T10:09:30.811 INFO:teuthology.orchestra.run.vm10.stdout:/dev/vdb 2026-03-07T10:09:30.811 INFO:teuthology.orchestra.run.vm10.stdout:/dev/vdc 2026-03-07T10:09:30.811 INFO:teuthology.orchestra.run.vm10.stdout:/dev/vdd 2026-03-07T10:09:30.811 INFO:teuthology.orchestra.run.vm10.stdout:/dev/vde 2026-03-07T10:09:30.811 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-07T10:09:30.811 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-07T10:09:30.811 DEBUG:teuthology.orchestra.run.vm10:> stat /dev/vdb 2026-03-07T10:09:30.854 INFO:teuthology.orchestra.run.vm10.stdout: File: /dev/vdb 2026-03-07T10:09:30.854 INFO:teuthology.orchestra.run.vm10.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-07T10:09:30.854 INFO:teuthology.orchestra.run.vm10.stdout:Device: 5h/5d Inode: 24 Links: 1 Device type: fe,10 2026-03-07T10:09:30.854 INFO:teuthology.orchestra.run.vm10.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-07T10:09:30.854 INFO:teuthology.orchestra.run.vm10.stdout:Access: 2026-03-07 10:04:04.981018550 +0000 2026-03-07T10:09:30.854 INFO:teuthology.orchestra.run.vm10.stdout:Modify: 2026-03-07 10:04:04.109018550 +0000 2026-03-07T10:09:30.854 INFO:teuthology.orchestra.run.vm10.stdout:Change: 2026-03-07 10:04:04.109018550 +0000 2026-03-07T10:09:30.854 INFO:teuthology.orchestra.run.vm10.stdout: Birth: - 2026-03-07T10:09:30.854 DEBUG:teuthology.orchestra.run.vm10:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-07T10:09:30.866 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:30 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:09:30.866 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:30 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:09:30.867 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:30 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:09:30.867 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:30 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:09:30.867 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:30 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:09:30.867 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:30 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:09:30.867 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:30 vm10 systemd[1]: Started Ceph mgr.b for 6c715b7a-1a0d-11f1-b180-89615ccd948e. 2026-03-07T10:09:30.877 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records in 2026-03-07T10:09:30.877 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records out 2026-03-07T10:09:30.877 INFO:teuthology.orchestra.run.vm10.stderr:512 bytes copied, 0.000102833 s, 5.0 MB/s 2026-03-07T10:09:30.878 DEBUG:teuthology.orchestra.run.vm10:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-07T10:09:30.930 DEBUG:teuthology.orchestra.run.vm10:> stat /dev/vdc 2026-03-07T10:09:30.977 INFO:teuthology.orchestra.run.vm10.stdout: File: /dev/vdc 2026-03-07T10:09:30.978 INFO:teuthology.orchestra.run.vm10.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-07T10:09:30.978 INFO:teuthology.orchestra.run.vm10.stdout:Device: 5h/5d Inode: 25 Links: 1 Device type: fe,20 2026-03-07T10:09:30.978 INFO:teuthology.orchestra.run.vm10.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-07T10:09:30.978 INFO:teuthology.orchestra.run.vm10.stdout:Access: 2026-03-07 10:04:04.989018550 +0000 2026-03-07T10:09:30.978 INFO:teuthology.orchestra.run.vm10.stdout:Modify: 2026-03-07 10:04:04.105018550 +0000 2026-03-07T10:09:30.978 INFO:teuthology.orchestra.run.vm10.stdout:Change: 2026-03-07 10:04:04.105018550 +0000 2026-03-07T10:09:30.978 INFO:teuthology.orchestra.run.vm10.stdout: Birth: - 2026-03-07T10:09:30.980 DEBUG:teuthology.orchestra.run.vm10:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-07T10:09:31.040 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records in 2026-03-07T10:09:31.040 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records out 2026-03-07T10:09:31.040 INFO:teuthology.orchestra.run.vm10.stderr:512 bytes copied, 0.00660784 s, 77.5 kB/s 2026-03-07T10:09:31.041 DEBUG:teuthology.orchestra.run.vm10:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-07T10:09:31.097 DEBUG:teuthology.orchestra.run.vm10:> stat /dev/vdd 2026-03-07T10:09:31.149 INFO:teuthology.orchestra.run.vm10.stdout: File: /dev/vdd 2026-03-07T10:09:31.149 INFO:teuthology.orchestra.run.vm10.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-07T10:09:31.150 INFO:teuthology.orchestra.run.vm10.stdout:Device: 5h/5d Inode: 26 Links: 1 Device type: fe,30 2026-03-07T10:09:31.150 INFO:teuthology.orchestra.run.vm10.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-07T10:09:31.150 INFO:teuthology.orchestra.run.vm10.stdout:Access: 2026-03-07 10:04:04.981018550 +0000 2026-03-07T10:09:31.150 INFO:teuthology.orchestra.run.vm10.stdout:Modify: 2026-03-07 10:04:04.121018550 +0000 2026-03-07T10:09:31.150 INFO:teuthology.orchestra.run.vm10.stdout:Change: 2026-03-07 10:04:04.121018550 +0000 2026-03-07T10:09:31.150 INFO:teuthology.orchestra.run.vm10.stdout: Birth: - 2026-03-07T10:09:31.150 DEBUG:teuthology.orchestra.run.vm10:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-07T10:09:31.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.084242+0000 mgr.a (mgr.14156) 51 : audit [DBG] from='client.24103 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm02=a;vm10=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:09:31.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.084242+0000 mgr.a (mgr.14156) 51 : audit [DBG] from='client.24103 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm02=a;vm10=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:09:31.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: cephadm 2026-03-07T10:09:30.085040+0000 mgr.a (mgr.14156) 52 : cephadm [INF] Saving service mgr spec with placement vm02=a;vm10=b;count:2 2026-03-07T10:09:31.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: cephadm 2026-03-07T10:09:30.085040+0000 mgr.a (mgr.14156) 52 : cephadm [INF] Saving service mgr spec with placement vm02=a;vm10=b;count:2 2026-03-07T10:09:31.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.088760+0000 mon.a (mon.0) 197 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:31.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.088760+0000 mon.a (mon.0) 197 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:31.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.089197+0000 mon.a (mon.0) 198 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:09:31.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.089197+0000 mon.a (mon.0) 198 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:09:31.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.090051+0000 mon.a (mon.0) 199 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:31.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.090051+0000 mon.a (mon.0) 199 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:31.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.090473+0000 mon.a (mon.0) 200 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:09:31.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.090473+0000 mon.a (mon.0) 200 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:09:31.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.093256+0000 mon.a (mon.0) 201 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:31.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.093256+0000 mon.a (mon.0) 201 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:31.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.093908+0000 mon.a (mon.0) 202 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-07T10:09:31.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.093908+0000 mon.a (mon.0) 202 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-07T10:09:31.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.095499+0000 mon.a (mon.0) 203 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-07T10:09:31.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.095499+0000 mon.a (mon.0) 203 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-07T10:09:31.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.096470+0000 mon.a (mon.0) 204 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-07T10:09:31.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.096470+0000 mon.a (mon.0) 204 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-07T10:09:31.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.096931+0000 mon.a (mon.0) 205 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:31.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.096931+0000 mon.a (mon.0) 205 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:31.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: cephadm 2026-03-07T10:09:30.097378+0000 mgr.a (mgr.14156) 53 : cephadm [INF] Deploying daemon mgr.b on vm10 2026-03-07T10:09:31.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: cephadm 2026-03-07T10:09:30.097378+0000 mgr.a (mgr.14156) 53 : cephadm [INF] Deploying daemon mgr.b on vm10 2026-03-07T10:09:31.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: cluster 2026-03-07T10:09:30.272177+0000 mgr.a (mgr.14156) 54 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:31.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: cluster 2026-03-07T10:09:30.272177+0000 mgr.a (mgr.14156) 54 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:31.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.896028+0000 mon.a (mon.0) 206 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:31.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.896028+0000 mon.a (mon.0) 206 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:31.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.899207+0000 mon.a (mon.0) 207 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:31.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.899207+0000 mon.a (mon.0) 207 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:31.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.902541+0000 mon.a (mon.0) 208 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:31.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.902541+0000 mon.a (mon.0) 208 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:31.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.906163+0000 mon.a (mon.0) 209 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:31.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.906163+0000 mon.a (mon.0) 209 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:31.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.916449+0000 mon.a (mon.0) 210 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:09:31.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19283]: audit 2026-03-07T10:09:30.916449+0000 mon.a (mon.0) 210 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:09:31.208 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records in 2026-03-07T10:09:31.208 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records out 2026-03-07T10:09:31.208 INFO:teuthology.orchestra.run.vm10.stderr:512 bytes copied, 0.00479468 s, 107 kB/s 2026-03-07T10:09:31.209 DEBUG:teuthology.orchestra.run.vm10:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-07T10:09:31.260 DEBUG:teuthology.orchestra.run.vm10:> stat /dev/vde 2026-03-07T10:09:31.307 INFO:teuthology.orchestra.run.vm10.stdout: File: /dev/vde 2026-03-07T10:09:31.307 INFO:teuthology.orchestra.run.vm10.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-07T10:09:31.307 INFO:teuthology.orchestra.run.vm10.stdout:Device: 5h/5d Inode: 27 Links: 1 Device type: fe,40 2026-03-07T10:09:31.307 INFO:teuthology.orchestra.run.vm10.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-07T10:09:31.307 INFO:teuthology.orchestra.run.vm10.stdout:Access: 2026-03-07 10:04:04.989018550 +0000 2026-03-07T10:09:31.307 INFO:teuthology.orchestra.run.vm10.stdout:Modify: 2026-03-07 10:04:04.109018550 +0000 2026-03-07T10:09:31.307 INFO:teuthology.orchestra.run.vm10.stdout:Change: 2026-03-07 10:04:04.109018550 +0000 2026-03-07T10:09:31.307 INFO:teuthology.orchestra.run.vm10.stdout: Birth: - 2026-03-07T10:09:31.307 DEBUG:teuthology.orchestra.run.vm10:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-07T10:09:31.354 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records in 2026-03-07T10:09:31.354 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records out 2026-03-07T10:09:31.354 INFO:teuthology.orchestra.run.vm10.stderr:512 bytes copied, 0.000192139 s, 2.7 MB/s 2026-03-07T10:09:31.355 DEBUG:teuthology.orchestra.run.vm10:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-07T10:09:31.401 INFO:tasks.cephadm:Deploying osd.0 on vm02 with /dev/vde... 2026-03-07T10:09:31.401 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- lvm zap /dev/vde 2026-03-07T10:09:31.408 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.084242+0000 mgr.a (mgr.14156) 51 : audit [DBG] from='client.24103 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm02=a;vm10=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:09:31.408 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.084242+0000 mgr.a (mgr.14156) 51 : audit [DBG] from='client.24103 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm02=a;vm10=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:09:31.408 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: cephadm 2026-03-07T10:09:30.085040+0000 mgr.a (mgr.14156) 52 : cephadm [INF] Saving service mgr spec with placement vm02=a;vm10=b;count:2 2026-03-07T10:09:31.408 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: cephadm 2026-03-07T10:09:30.085040+0000 mgr.a (mgr.14156) 52 : cephadm [INF] Saving service mgr spec with placement vm02=a;vm10=b;count:2 2026-03-07T10:09:31.408 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.088760+0000 mon.a (mon.0) 197 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:31.408 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.088760+0000 mon.a (mon.0) 197 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:31.408 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.089197+0000 mon.a (mon.0) 198 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:09:31.408 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.089197+0000 mon.a (mon.0) 198 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:09:31.408 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.090051+0000 mon.a (mon.0) 199 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:31.408 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.090051+0000 mon.a (mon.0) 199 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:31.408 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.090473+0000 mon.a (mon.0) 200 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:09:31.408 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.090473+0000 mon.a (mon.0) 200 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:09:31.408 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.093256+0000 mon.a (mon.0) 201 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:31.408 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.093256+0000 mon.a (mon.0) 201 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:31.408 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.093908+0000 mon.a (mon.0) 202 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-07T10:09:31.408 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.093908+0000 mon.a (mon.0) 202 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-07T10:09:31.408 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.095499+0000 mon.a (mon.0) 203 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-07T10:09:31.408 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.095499+0000 mon.a (mon.0) 203 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-07T10:09:31.408 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.096470+0000 mon.a (mon.0) 204 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-07T10:09:31.408 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.096470+0000 mon.a (mon.0) 204 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-07T10:09:31.408 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.096931+0000 mon.a (mon.0) 205 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:31.409 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.096931+0000 mon.a (mon.0) 205 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:31.409 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: cephadm 2026-03-07T10:09:30.097378+0000 mgr.a (mgr.14156) 53 : cephadm [INF] Deploying daemon mgr.b on vm10 2026-03-07T10:09:31.409 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: cephadm 2026-03-07T10:09:30.097378+0000 mgr.a (mgr.14156) 53 : cephadm [INF] Deploying daemon mgr.b on vm10 2026-03-07T10:09:31.409 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: cluster 2026-03-07T10:09:30.272177+0000 mgr.a (mgr.14156) 54 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:31.409 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: cluster 2026-03-07T10:09:30.272177+0000 mgr.a (mgr.14156) 54 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:31.409 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.896028+0000 mon.a (mon.0) 206 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:31.409 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.896028+0000 mon.a (mon.0) 206 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:31.409 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.899207+0000 mon.a (mon.0) 207 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:31.409 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.899207+0000 mon.a (mon.0) 207 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:31.409 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.902541+0000 mon.a (mon.0) 208 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:31.409 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.902541+0000 mon.a (mon.0) 208 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:31.409 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.906163+0000 mon.a (mon.0) 209 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:31.409 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.906163+0000 mon.a (mon.0) 209 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:31.409 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.916449+0000 mon.a (mon.0) 210 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:09:31.409 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:31 vm02 bash[17011]: audit 2026-03-07T10:09:30.916449+0000 mon.a (mon.0) 210 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:09:31.695 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19881]: debug 2026-03-07T10:09:31.272+0000 7fb2c2ec0100 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-07T10:09:31.695 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19881]: debug 2026-03-07T10:09:31.404+0000 7fb2c2ec0100 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-07T10:09:32.195 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:31 vm10 bash[19881]: debug 2026-03-07T10:09:31.780+0000 7fb2c2ec0100 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-07T10:09:33.195 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:32 vm10 bash[19881]: debug 2026-03-07T10:09:32.896+0000 7fb2c2ec0100 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-07T10:09:33.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:33 vm10 bash[19283]: cluster 2026-03-07T10:09:32.272387+0000 mgr.a (mgr.14156) 55 : cluster [DBG] pgmap v22: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:33.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:33 vm10 bash[19283]: cluster 2026-03-07T10:09:32.272387+0000 mgr.a (mgr.14156) 55 : cluster [DBG] pgmap v22: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:33.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:33 vm02 bash[17011]: cluster 2026-03-07T10:09:32.272387+0000 mgr.a (mgr.14156) 55 : cluster [DBG] pgmap v22: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:33.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:33 vm02 bash[17011]: cluster 2026-03-07T10:09:32.272387+0000 mgr.a (mgr.14156) 55 : cluster [DBG] pgmap v22: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:34.549 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:34 vm10 bash[19881]: debug 2026-03-07T10:09:34.228+0000 7fb2c2ec0100 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-07T10:09:34.945 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:34 vm10 bash[19881]: debug 2026-03-07T10:09:34.548+0000 7fb2c2ec0100 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-07T10:09:35.331 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:35 vm10 bash[19881]: debug 2026-03-07T10:09:35.064+0000 7fb2c2ec0100 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-07T10:09:35.331 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:35 vm10 bash[19881]: debug 2026-03-07T10:09:35.192+0000 7fb2c2ec0100 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-07T10:09:35.664 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:35 vm10 bash[19283]: cluster 2026-03-07T10:09:34.272586+0000 mgr.a (mgr.14156) 56 : cluster [DBG] pgmap v23: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:35.664 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:35 vm10 bash[19283]: cluster 2026-03-07T10:09:34.272586+0000 mgr.a (mgr.14156) 56 : cluster [DBG] pgmap v23: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:35.664 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:35 vm10 bash[19283]: audit 2026-03-07T10:09:34.344545+0000 mon.a (mon.0) 211 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:35.664 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:35 vm10 bash[19283]: audit 2026-03-07T10:09:34.344545+0000 mon.a (mon.0) 211 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:35.664 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:35 vm10 bash[19881]: debug 2026-03-07T10:09:35.328+0000 7fb2c2ec0100 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-07T10:09:35.664 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:35 vm10 bash[19881]: debug 2026-03-07T10:09:35.492+0000 7fb2c2ec0100 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-07T10:09:35.664 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:35 vm10 bash[19881]: debug 2026-03-07T10:09:35.664+0000 7fb2c2ec0100 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-07T10:09:35.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:35 vm02 bash[17011]: cluster 2026-03-07T10:09:34.272586+0000 mgr.a (mgr.14156) 56 : cluster [DBG] pgmap v23: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:35.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:35 vm02 bash[17011]: cluster 2026-03-07T10:09:34.272586+0000 mgr.a (mgr.14156) 56 : cluster [DBG] pgmap v23: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:35.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:35 vm02 bash[17011]: audit 2026-03-07T10:09:34.344545+0000 mon.a (mon.0) 211 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:35.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:35 vm02 bash[17011]: audit 2026-03-07T10:09:34.344545+0000 mon.a (mon.0) 211 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:36.183 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config 2026-03-07T10:09:37.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: audit 2026-03-07T10:09:35.915016+0000 mon.a (mon.0) 212 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: audit 2026-03-07T10:09:35.915016+0000 mon.a (mon.0) 212 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: audit 2026-03-07T10:09:35.918483+0000 mon.a (mon.0) 213 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: audit 2026-03-07T10:09:35.918483+0000 mon.a (mon.0) 213 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: audit 2026-03-07T10:09:35.919462+0000 mon.a (mon.0) 214 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: audit 2026-03-07T10:09:35.919462+0000 mon.a (mon.0) 214 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: audit 2026-03-07T10:09:35.919891+0000 mon.a (mon.0) 215 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: audit 2026-03-07T10:09:35.919891+0000 mon.a (mon.0) 215 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: audit 2026-03-07T10:09:35.922880+0000 mon.a (mon.0) 216 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: audit 2026-03-07T10:09:35.922880+0000 mon.a (mon.0) 216 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: cephadm 2026-03-07T10:09:35.932387+0000 mgr.a (mgr.14156) 57 : cephadm [INF] Reconfiguring mgr.a (unknown last config time)... 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: cephadm 2026-03-07T10:09:35.932387+0000 mgr.a (mgr.14156) 57 : cephadm [INF] Reconfiguring mgr.a (unknown last config time)... 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: audit 2026-03-07T10:09:35.932549+0000 mon.a (mon.0) 217 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: audit 2026-03-07T10:09:35.932549+0000 mon.a (mon.0) 217 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: audit 2026-03-07T10:09:35.932985+0000 mon.a (mon.0) 218 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: audit 2026-03-07T10:09:35.932985+0000 mon.a (mon.0) 218 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: audit 2026-03-07T10:09:35.933353+0000 mon.a (mon.0) 219 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: audit 2026-03-07T10:09:35.933353+0000 mon.a (mon.0) 219 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: cephadm 2026-03-07T10:09:35.933740+0000 mgr.a (mgr.14156) 58 : cephadm [INF] Reconfiguring daemon mgr.a on vm02 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: cephadm 2026-03-07T10:09:35.933740+0000 mgr.a (mgr.14156) 58 : cephadm [INF] Reconfiguring daemon mgr.a on vm02 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: audit 2026-03-07T10:09:36.344317+0000 mon.a (mon.0) 220 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: audit 2026-03-07T10:09:36.344317+0000 mon.a (mon.0) 220 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: audit 2026-03-07T10:09:36.348800+0000 mon.a (mon.0) 221 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: audit 2026-03-07T10:09:36.348800+0000 mon.a (mon.0) 221 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: audit 2026-03-07T10:09:36.350053+0000 mon.a (mon.0) 222 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: audit 2026-03-07T10:09:36.350053+0000 mon.a (mon.0) 222 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: audit 2026-03-07T10:09:36.351002+0000 mon.a (mon.0) 223 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: audit 2026-03-07T10:09:36.351002+0000 mon.a (mon.0) 223 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: audit 2026-03-07T10:09:36.351359+0000 mon.a (mon.0) 224 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: audit 2026-03-07T10:09:36.351359+0000 mon.a (mon.0) 224 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: audit 2026-03-07T10:09:36.354927+0000 mon.a (mon.0) 225 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:36 vm10 bash[19283]: audit 2026-03-07T10:09:36.354927+0000 mon.a (mon.0) 225 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:37 vm10 bash[19881]: debug 2026-03-07T10:09:37.012+0000 7fb2c2ec0100 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-07T10:09:37.196 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:37 vm10 bash[19881]: debug 2026-03-07T10:09:37.148+0000 7fb2c2ec0100 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: audit 2026-03-07T10:09:35.915016+0000 mon.a (mon.0) 212 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: audit 2026-03-07T10:09:35.915016+0000 mon.a (mon.0) 212 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: audit 2026-03-07T10:09:35.918483+0000 mon.a (mon.0) 213 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: audit 2026-03-07T10:09:35.918483+0000 mon.a (mon.0) 213 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: audit 2026-03-07T10:09:35.919462+0000 mon.a (mon.0) 214 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: audit 2026-03-07T10:09:35.919462+0000 mon.a (mon.0) 214 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: audit 2026-03-07T10:09:35.919891+0000 mon.a (mon.0) 215 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: audit 2026-03-07T10:09:35.919891+0000 mon.a (mon.0) 215 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: audit 2026-03-07T10:09:35.922880+0000 mon.a (mon.0) 216 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: audit 2026-03-07T10:09:35.922880+0000 mon.a (mon.0) 216 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: cephadm 2026-03-07T10:09:35.932387+0000 mgr.a (mgr.14156) 57 : cephadm [INF] Reconfiguring mgr.a (unknown last config time)... 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: cephadm 2026-03-07T10:09:35.932387+0000 mgr.a (mgr.14156) 57 : cephadm [INF] Reconfiguring mgr.a (unknown last config time)... 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: audit 2026-03-07T10:09:35.932549+0000 mon.a (mon.0) 217 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: audit 2026-03-07T10:09:35.932549+0000 mon.a (mon.0) 217 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: audit 2026-03-07T10:09:35.932985+0000 mon.a (mon.0) 218 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: audit 2026-03-07T10:09:35.932985+0000 mon.a (mon.0) 218 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: audit 2026-03-07T10:09:35.933353+0000 mon.a (mon.0) 219 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: audit 2026-03-07T10:09:35.933353+0000 mon.a (mon.0) 219 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: cephadm 2026-03-07T10:09:35.933740+0000 mgr.a (mgr.14156) 58 : cephadm [INF] Reconfiguring daemon mgr.a on vm02 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: cephadm 2026-03-07T10:09:35.933740+0000 mgr.a (mgr.14156) 58 : cephadm [INF] Reconfiguring daemon mgr.a on vm02 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: audit 2026-03-07T10:09:36.344317+0000 mon.a (mon.0) 220 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: audit 2026-03-07T10:09:36.344317+0000 mon.a (mon.0) 220 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: audit 2026-03-07T10:09:36.348800+0000 mon.a (mon.0) 221 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: audit 2026-03-07T10:09:36.348800+0000 mon.a (mon.0) 221 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: audit 2026-03-07T10:09:36.350053+0000 mon.a (mon.0) 222 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: audit 2026-03-07T10:09:36.350053+0000 mon.a (mon.0) 222 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: audit 2026-03-07T10:09:36.351002+0000 mon.a (mon.0) 223 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: audit 2026-03-07T10:09:36.351002+0000 mon.a (mon.0) 223 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: audit 2026-03-07T10:09:36.351359+0000 mon.a (mon.0) 224 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: audit 2026-03-07T10:09:36.351359+0000 mon.a (mon.0) 224 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: audit 2026-03-07T10:09:36.354927+0000 mon.a (mon.0) 225 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:36 vm02 bash[17011]: audit 2026-03-07T10:09:36.354927+0000 mon.a (mon.0) 225 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:09:37.417 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-07T10:09:37.435 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph orch daemon add osd vm02:/dev/vde 2026-03-07T10:09:37.695 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:37 vm10 bash[19881]: debug 2026-03-07T10:09:37.276+0000 7fb2c2ec0100 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-07T10:09:38.133 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:37 vm10 bash[19283]: cluster 2026-03-07T10:09:36.272771+0000 mgr.a (mgr.14156) 59 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:38.133 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:37 vm10 bash[19283]: cluster 2026-03-07T10:09:36.272771+0000 mgr.a (mgr.14156) 59 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:38.133 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:37 vm10 bash[19881]: debug 2026-03-07T10:09:37.832+0000 7fb2c2ec0100 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-07T10:09:38.133 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:37 vm10 bash[19881]: debug 2026-03-07T10:09:37.980+0000 7fb2c2ec0100 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-07T10:09:38.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:37 vm02 bash[17011]: cluster 2026-03-07T10:09:36.272771+0000 mgr.a (mgr.14156) 59 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:38.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:37 vm02 bash[17011]: cluster 2026-03-07T10:09:36.272771+0000 mgr.a (mgr.14156) 59 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:38.445 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:38 vm10 bash[19881]: debug 2026-03-07T10:09:38.132+0000 7fb2c2ec0100 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-07T10:09:38.945 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:38 vm10 bash[19881]: debug 2026-03-07T10:09:38.484+0000 7fb2c2ec0100 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-07T10:09:39.445 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:38 vm10 bash[19881]: debug 2026-03-07T10:09:38.996+0000 7fb2c2ec0100 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-07T10:09:39.807 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:39 vm10 bash[19881]: debug 2026-03-07T10:09:39.532+0000 7fb2c2ec0100 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-07T10:09:39.807 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:39 vm10 bash[19881]: debug 2026-03-07T10:09:39.672+0000 7fb2c2ec0100 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-07T10:09:40.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:39 vm10 bash[19283]: cluster 2026-03-07T10:09:38.272973+0000 mgr.a (mgr.14156) 60 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:40.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:39 vm10 bash[19283]: cluster 2026-03-07T10:09:38.272973+0000 mgr.a (mgr.14156) 60 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:40.196 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:39 vm10 bash[19881]: debug 2026-03-07T10:09:39.804+0000 7fb2c2ec0100 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-07T10:09:40.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:39 vm02 bash[17011]: cluster 2026-03-07T10:09:38.272973+0000 mgr.a (mgr.14156) 60 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:40.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:39 vm02 bash[17011]: cluster 2026-03-07T10:09:38.272973+0000 mgr.a (mgr.14156) 60 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:40.695 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:40 vm10 bash[19881]: debug 2026-03-07T10:09:40.280+0000 7fb2c2ec0100 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-07T10:09:41.360 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:41 vm10 bash[19283]: cluster 2026-03-07T10:09:40.273179+0000 mgr.a (mgr.14156) 61 : cluster [DBG] pgmap v26: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:41.360 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:41 vm10 bash[19283]: cluster 2026-03-07T10:09:40.273179+0000 mgr.a (mgr.14156) 61 : cluster [DBG] pgmap v26: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:41.461 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:41 vm02 bash[17011]: cluster 2026-03-07T10:09:40.273179+0000 mgr.a (mgr.14156) 61 : cluster [DBG] pgmap v26: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:41.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:41 vm02 bash[17011]: cluster 2026-03-07T10:09:40.273179+0000 mgr.a (mgr.14156) 61 : cluster [DBG] pgmap v26: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:41.695 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:09:41 vm10 bash[19881]: debug 2026-03-07T10:09:41.356+0000 7fb2c2ec0100 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-07T10:09:42.235 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config 2026-03-07T10:09:42.445 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:42 vm10 bash[19283]: cluster 2026-03-07T10:09:41.361901+0000 mon.a (mon.0) 226 : cluster [DBG] Standby manager daemon b started 2026-03-07T10:09:42.446 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:42 vm10 bash[19283]: cluster 2026-03-07T10:09:41.361901+0000 mon.a (mon.0) 226 : cluster [DBG] Standby manager daemon b started 2026-03-07T10:09:42.446 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:42 vm10 bash[19283]: audit 2026-03-07T10:09:41.363233+0000 mon.b (mon.1) 2 : audit [DBG] from='mgr.? 192.168.123.110:0/3302515155' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-07T10:09:42.446 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:42 vm10 bash[19283]: audit 2026-03-07T10:09:41.363233+0000 mon.b (mon.1) 2 : audit [DBG] from='mgr.? 192.168.123.110:0/3302515155' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-07T10:09:42.446 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:42 vm10 bash[19283]: audit 2026-03-07T10:09:41.363855+0000 mon.b (mon.1) 3 : audit [DBG] from='mgr.? 192.168.123.110:0/3302515155' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-07T10:09:42.446 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:42 vm10 bash[19283]: audit 2026-03-07T10:09:41.363855+0000 mon.b (mon.1) 3 : audit [DBG] from='mgr.? 192.168.123.110:0/3302515155' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-07T10:09:42.446 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:42 vm10 bash[19283]: audit 2026-03-07T10:09:41.364620+0000 mon.b (mon.1) 4 : audit [DBG] from='mgr.? 192.168.123.110:0/3302515155' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-07T10:09:42.446 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:42 vm10 bash[19283]: audit 2026-03-07T10:09:41.364620+0000 mon.b (mon.1) 4 : audit [DBG] from='mgr.? 192.168.123.110:0/3302515155' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-07T10:09:42.446 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:42 vm10 bash[19283]: audit 2026-03-07T10:09:41.365010+0000 mon.b (mon.1) 5 : audit [DBG] from='mgr.? 192.168.123.110:0/3302515155' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-07T10:09:42.446 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:42 vm10 bash[19283]: audit 2026-03-07T10:09:41.365010+0000 mon.b (mon.1) 5 : audit [DBG] from='mgr.? 192.168.123.110:0/3302515155' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-07T10:09:42.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:42 vm02 bash[17011]: cluster 2026-03-07T10:09:41.361901+0000 mon.a (mon.0) 226 : cluster [DBG] Standby manager daemon b started 2026-03-07T10:09:42.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:42 vm02 bash[17011]: cluster 2026-03-07T10:09:41.361901+0000 mon.a (mon.0) 226 : cluster [DBG] Standby manager daemon b started 2026-03-07T10:09:42.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:42 vm02 bash[17011]: audit 2026-03-07T10:09:41.363233+0000 mon.b (mon.1) 2 : audit [DBG] from='mgr.? 192.168.123.110:0/3302515155' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-07T10:09:42.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:42 vm02 bash[17011]: audit 2026-03-07T10:09:41.363233+0000 mon.b (mon.1) 2 : audit [DBG] from='mgr.? 192.168.123.110:0/3302515155' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-07T10:09:42.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:42 vm02 bash[17011]: audit 2026-03-07T10:09:41.363855+0000 mon.b (mon.1) 3 : audit [DBG] from='mgr.? 192.168.123.110:0/3302515155' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-07T10:09:42.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:42 vm02 bash[17011]: audit 2026-03-07T10:09:41.363855+0000 mon.b (mon.1) 3 : audit [DBG] from='mgr.? 192.168.123.110:0/3302515155' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-07T10:09:42.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:42 vm02 bash[17011]: audit 2026-03-07T10:09:41.364620+0000 mon.b (mon.1) 4 : audit [DBG] from='mgr.? 192.168.123.110:0/3302515155' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-07T10:09:42.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:42 vm02 bash[17011]: audit 2026-03-07T10:09:41.364620+0000 mon.b (mon.1) 4 : audit [DBG] from='mgr.? 192.168.123.110:0/3302515155' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-07T10:09:42.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:42 vm02 bash[17011]: audit 2026-03-07T10:09:41.365010+0000 mon.b (mon.1) 5 : audit [DBG] from='mgr.? 192.168.123.110:0/3302515155' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-07T10:09:42.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:42 vm02 bash[17011]: audit 2026-03-07T10:09:41.365010+0000 mon.b (mon.1) 5 : audit [DBG] from='mgr.? 192.168.123.110:0/3302515155' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-07T10:09:43.556 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:43 vm10 bash[19283]: cluster 2026-03-07T10:09:42.101161+0000 mon.a (mon.0) 227 : cluster [DBG] mgrmap e14: a(active, since 65s), standbys: b 2026-03-07T10:09:43.556 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:43 vm10 bash[19283]: cluster 2026-03-07T10:09:42.101161+0000 mon.a (mon.0) 227 : cluster [DBG] mgrmap e14: a(active, since 65s), standbys: b 2026-03-07T10:09:43.556 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:43 vm10 bash[19283]: audit 2026-03-07T10:09:42.101249+0000 mon.a (mon.0) 228 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-07T10:09:43.556 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:43 vm10 bash[19283]: audit 2026-03-07T10:09:42.101249+0000 mon.a (mon.0) 228 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-07T10:09:43.556 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:43 vm10 bash[19283]: cluster 2026-03-07T10:09:42.273351+0000 mgr.a (mgr.14156) 62 : cluster [DBG] pgmap v27: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:43.556 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:43 vm10 bash[19283]: cluster 2026-03-07T10:09:42.273351+0000 mgr.a (mgr.14156) 62 : cluster [DBG] pgmap v27: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:43.556 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:43 vm10 bash[19283]: audit 2026-03-07T10:09:42.603771+0000 mgr.a (mgr.14156) 63 : audit [DBG] from='client.14200 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:09:43.556 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:43 vm10 bash[19283]: audit 2026-03-07T10:09:42.603771+0000 mgr.a (mgr.14156) 63 : audit [DBG] from='client.14200 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:09:43.556 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:43 vm10 bash[19283]: audit 2026-03-07T10:09:42.605264+0000 mon.a (mon.0) 229 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T10:09:43.556 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:43 vm10 bash[19283]: audit 2026-03-07T10:09:42.605264+0000 mon.a (mon.0) 229 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T10:09:43.556 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:43 vm10 bash[19283]: audit 2026-03-07T10:09:42.607525+0000 mon.a (mon.0) 230 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-07T10:09:43.556 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:43 vm10 bash[19283]: audit 2026-03-07T10:09:42.607525+0000 mon.a (mon.0) 230 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-07T10:09:43.556 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:43 vm10 bash[19283]: audit 2026-03-07T10:09:42.608479+0000 mon.a (mon.0) 231 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:43.556 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:43 vm10 bash[19283]: audit 2026-03-07T10:09:42.608479+0000 mon.a (mon.0) 231 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:43.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:43 vm02 bash[17011]: cluster 2026-03-07T10:09:42.101161+0000 mon.a (mon.0) 227 : cluster [DBG] mgrmap e14: a(active, since 65s), standbys: b 2026-03-07T10:09:43.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:43 vm02 bash[17011]: cluster 2026-03-07T10:09:42.101161+0000 mon.a (mon.0) 227 : cluster [DBG] mgrmap e14: a(active, since 65s), standbys: b 2026-03-07T10:09:43.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:43 vm02 bash[17011]: audit 2026-03-07T10:09:42.101249+0000 mon.a (mon.0) 228 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-07T10:09:43.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:43 vm02 bash[17011]: audit 2026-03-07T10:09:42.101249+0000 mon.a (mon.0) 228 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-07T10:09:43.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:43 vm02 bash[17011]: cluster 2026-03-07T10:09:42.273351+0000 mgr.a (mgr.14156) 62 : cluster [DBG] pgmap v27: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:43.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:43 vm02 bash[17011]: cluster 2026-03-07T10:09:42.273351+0000 mgr.a (mgr.14156) 62 : cluster [DBG] pgmap v27: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:43.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:43 vm02 bash[17011]: audit 2026-03-07T10:09:42.603771+0000 mgr.a (mgr.14156) 63 : audit [DBG] from='client.14200 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:09:43.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:43 vm02 bash[17011]: audit 2026-03-07T10:09:42.603771+0000 mgr.a (mgr.14156) 63 : audit [DBG] from='client.14200 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:09:43.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:43 vm02 bash[17011]: audit 2026-03-07T10:09:42.605264+0000 mon.a (mon.0) 229 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T10:09:43.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:43 vm02 bash[17011]: audit 2026-03-07T10:09:42.605264+0000 mon.a (mon.0) 229 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T10:09:43.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:43 vm02 bash[17011]: audit 2026-03-07T10:09:42.607525+0000 mon.a (mon.0) 230 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-07T10:09:43.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:43 vm02 bash[17011]: audit 2026-03-07T10:09:42.607525+0000 mon.a (mon.0) 230 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-07T10:09:43.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:43 vm02 bash[17011]: audit 2026-03-07T10:09:42.608479+0000 mon.a (mon.0) 231 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:43.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:43 vm02 bash[17011]: audit 2026-03-07T10:09:42.608479+0000 mon.a (mon.0) 231 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:45.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:45 vm10 bash[19283]: cluster 2026-03-07T10:09:44.273511+0000 mgr.a (mgr.14156) 64 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:45.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:45 vm10 bash[19283]: cluster 2026-03-07T10:09:44.273511+0000 mgr.a (mgr.14156) 64 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:45.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:45 vm02 bash[17011]: cluster 2026-03-07T10:09:44.273511+0000 mgr.a (mgr.14156) 64 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:45.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:45 vm02 bash[17011]: cluster 2026-03-07T10:09:44.273511+0000 mgr.a (mgr.14156) 64 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:47.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:47 vm10 bash[19283]: cluster 2026-03-07T10:09:46.273686+0000 mgr.a (mgr.14156) 65 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:47.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:47 vm10 bash[19283]: cluster 2026-03-07T10:09:46.273686+0000 mgr.a (mgr.14156) 65 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:47.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:47 vm02 bash[17011]: cluster 2026-03-07T10:09:46.273686+0000 mgr.a (mgr.14156) 65 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:47.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:47 vm02 bash[17011]: cluster 2026-03-07T10:09:46.273686+0000 mgr.a (mgr.14156) 65 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:48.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:48 vm10 bash[19283]: audit 2026-03-07T10:09:48.277657+0000 mon.a (mon.0) 232 : audit [INF] from='client.? 192.168.123.102:0/3983940649' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "eb891bbe-0ffc-4a69-afbc-01132f549280"}]: dispatch 2026-03-07T10:09:48.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:48 vm10 bash[19283]: audit 2026-03-07T10:09:48.277657+0000 mon.a (mon.0) 232 : audit [INF] from='client.? 192.168.123.102:0/3983940649' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "eb891bbe-0ffc-4a69-afbc-01132f549280"}]: dispatch 2026-03-07T10:09:48.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:48 vm10 bash[19283]: audit 2026-03-07T10:09:48.281109+0000 mon.a (mon.0) 233 : audit [INF] from='client.? 192.168.123.102:0/3983940649' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "eb891bbe-0ffc-4a69-afbc-01132f549280"}]': finished 2026-03-07T10:09:48.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:48 vm10 bash[19283]: audit 2026-03-07T10:09:48.281109+0000 mon.a (mon.0) 233 : audit [INF] from='client.? 192.168.123.102:0/3983940649' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "eb891bbe-0ffc-4a69-afbc-01132f549280"}]': finished 2026-03-07T10:09:48.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:48 vm10 bash[19283]: cluster 2026-03-07T10:09:48.283396+0000 mon.a (mon.0) 234 : cluster [DBG] osdmap e5: 1 total, 0 up, 1 in 2026-03-07T10:09:48.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:48 vm10 bash[19283]: cluster 2026-03-07T10:09:48.283396+0000 mon.a (mon.0) 234 : cluster [DBG] osdmap e5: 1 total, 0 up, 1 in 2026-03-07T10:09:48.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:48 vm10 bash[19283]: audit 2026-03-07T10:09:48.283747+0000 mon.a (mon.0) 235 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T10:09:48.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:48 vm10 bash[19283]: audit 2026-03-07T10:09:48.283747+0000 mon.a (mon.0) 235 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T10:09:48.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:48 vm02 bash[17011]: audit 2026-03-07T10:09:48.277657+0000 mon.a (mon.0) 232 : audit [INF] from='client.? 192.168.123.102:0/3983940649' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "eb891bbe-0ffc-4a69-afbc-01132f549280"}]: dispatch 2026-03-07T10:09:48.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:48 vm02 bash[17011]: audit 2026-03-07T10:09:48.277657+0000 mon.a (mon.0) 232 : audit [INF] from='client.? 192.168.123.102:0/3983940649' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "eb891bbe-0ffc-4a69-afbc-01132f549280"}]: dispatch 2026-03-07T10:09:48.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:48 vm02 bash[17011]: audit 2026-03-07T10:09:48.281109+0000 mon.a (mon.0) 233 : audit [INF] from='client.? 192.168.123.102:0/3983940649' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "eb891bbe-0ffc-4a69-afbc-01132f549280"}]': finished 2026-03-07T10:09:48.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:48 vm02 bash[17011]: audit 2026-03-07T10:09:48.281109+0000 mon.a (mon.0) 233 : audit [INF] from='client.? 192.168.123.102:0/3983940649' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "eb891bbe-0ffc-4a69-afbc-01132f549280"}]': finished 2026-03-07T10:09:48.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:48 vm02 bash[17011]: cluster 2026-03-07T10:09:48.283396+0000 mon.a (mon.0) 234 : cluster [DBG] osdmap e5: 1 total, 0 up, 1 in 2026-03-07T10:09:48.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:48 vm02 bash[17011]: cluster 2026-03-07T10:09:48.283396+0000 mon.a (mon.0) 234 : cluster [DBG] osdmap e5: 1 total, 0 up, 1 in 2026-03-07T10:09:48.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:48 vm02 bash[17011]: audit 2026-03-07T10:09:48.283747+0000 mon.a (mon.0) 235 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T10:09:48.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:48 vm02 bash[17011]: audit 2026-03-07T10:09:48.283747+0000 mon.a (mon.0) 235 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T10:09:49.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:49 vm10 bash[19283]: cluster 2026-03-07T10:09:48.273854+0000 mgr.a (mgr.14156) 66 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:49.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:49 vm10 bash[19283]: cluster 2026-03-07T10:09:48.273854+0000 mgr.a (mgr.14156) 66 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:49.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:49 vm10 bash[19283]: audit 2026-03-07T10:09:48.879612+0000 mon.a (mon.0) 236 : audit [DBG] from='client.? 192.168.123.102:0/1237833710' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T10:09:49.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:49 vm10 bash[19283]: audit 2026-03-07T10:09:48.879612+0000 mon.a (mon.0) 236 : audit [DBG] from='client.? 192.168.123.102:0/1237833710' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T10:09:49.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:49 vm02 bash[17011]: cluster 2026-03-07T10:09:48.273854+0000 mgr.a (mgr.14156) 66 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:49.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:49 vm02 bash[17011]: cluster 2026-03-07T10:09:48.273854+0000 mgr.a (mgr.14156) 66 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:49.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:49 vm02 bash[17011]: audit 2026-03-07T10:09:48.879612+0000 mon.a (mon.0) 236 : audit [DBG] from='client.? 192.168.123.102:0/1237833710' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T10:09:49.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:49 vm02 bash[17011]: audit 2026-03-07T10:09:48.879612+0000 mon.a (mon.0) 236 : audit [DBG] from='client.? 192.168.123.102:0/1237833710' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T10:09:51.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:51 vm10 bash[19283]: cluster 2026-03-07T10:09:50.274018+0000 mgr.a (mgr.14156) 67 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:51.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:51 vm10 bash[19283]: cluster 2026-03-07T10:09:50.274018+0000 mgr.a (mgr.14156) 67 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:51.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:51 vm02 bash[17011]: cluster 2026-03-07T10:09:50.274018+0000 mgr.a (mgr.14156) 67 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:51.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:51 vm02 bash[17011]: cluster 2026-03-07T10:09:50.274018+0000 mgr.a (mgr.14156) 67 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:53.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:53 vm10 bash[19283]: cluster 2026-03-07T10:09:52.274193+0000 mgr.a (mgr.14156) 68 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:53.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:53 vm10 bash[19283]: cluster 2026-03-07T10:09:52.274193+0000 mgr.a (mgr.14156) 68 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:53.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:53 vm02 bash[17011]: cluster 2026-03-07T10:09:52.274193+0000 mgr.a (mgr.14156) 68 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:53.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:53 vm02 bash[17011]: cluster 2026-03-07T10:09:52.274193+0000 mgr.a (mgr.14156) 68 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:55.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:55 vm10 bash[19283]: cluster 2026-03-07T10:09:54.274352+0000 mgr.a (mgr.14156) 69 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:55.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:55 vm10 bash[19283]: cluster 2026-03-07T10:09:54.274352+0000 mgr.a (mgr.14156) 69 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:55.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:55 vm02 bash[17011]: cluster 2026-03-07T10:09:54.274352+0000 mgr.a (mgr.14156) 69 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:55.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:55 vm02 bash[17011]: cluster 2026-03-07T10:09:54.274352+0000 mgr.a (mgr.14156) 69 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:57.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:57 vm10 bash[19283]: cluster 2026-03-07T10:09:56.274499+0000 mgr.a (mgr.14156) 70 : cluster [DBG] pgmap v35: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:57.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:57 vm10 bash[19283]: cluster 2026-03-07T10:09:56.274499+0000 mgr.a (mgr.14156) 70 : cluster [DBG] pgmap v35: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:57.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:57 vm02 bash[17011]: cluster 2026-03-07T10:09:56.274499+0000 mgr.a (mgr.14156) 70 : cluster [DBG] pgmap v35: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:57.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:57 vm02 bash[17011]: cluster 2026-03-07T10:09:56.274499+0000 mgr.a (mgr.14156) 70 : cluster [DBG] pgmap v35: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:58.775 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:58 vm02 bash[17011]: audit 2026-03-07T10:09:58.251904+0000 mon.a (mon.0) 237 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-07T10:09:58.776 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:58 vm02 bash[17011]: audit 2026-03-07T10:09:58.251904+0000 mon.a (mon.0) 237 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-07T10:09:58.776 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:58 vm02 bash[17011]: audit 2026-03-07T10:09:58.252501+0000 mon.a (mon.0) 238 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:58.776 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:58 vm02 bash[17011]: audit 2026-03-07T10:09:58.252501+0000 mon.a (mon.0) 238 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:58.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:58 vm10 bash[19283]: audit 2026-03-07T10:09:58.251904+0000 mon.a (mon.0) 237 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-07T10:09:58.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:58 vm10 bash[19283]: audit 2026-03-07T10:09:58.251904+0000 mon.a (mon.0) 237 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-07T10:09:58.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:58 vm10 bash[19283]: audit 2026-03-07T10:09:58.252501+0000 mon.a (mon.0) 238 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:58.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:58 vm10 bash[19283]: audit 2026-03-07T10:09:58.252501+0000 mon.a (mon.0) 238 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:09:59.712 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:09:59 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:09:59.712 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:09:59 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:09:59.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:59 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:09:59.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:59 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:09:59.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:59 vm02 bash[17011]: cephadm 2026-03-07T10:09:58.252983+0000 mgr.a (mgr.14156) 71 : cephadm [INF] Deploying daemon osd.0 on vm02 2026-03-07T10:09:59.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:59 vm02 bash[17011]: cephadm 2026-03-07T10:09:58.252983+0000 mgr.a (mgr.14156) 71 : cephadm [INF] Deploying daemon osd.0 on vm02 2026-03-07T10:09:59.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:59 vm02 bash[17011]: cluster 2026-03-07T10:09:58.274680+0000 mgr.a (mgr.14156) 72 : cluster [DBG] pgmap v36: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:09:59.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:09:59 vm02 bash[17011]: cluster 2026-03-07T10:09:58.274680+0000 mgr.a (mgr.14156) 72 : cluster [DBG] pgmap v36: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:10:00.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:59 vm10 bash[19283]: cephadm 2026-03-07T10:09:58.252983+0000 mgr.a (mgr.14156) 71 : cephadm [INF] Deploying daemon osd.0 on vm02 2026-03-07T10:10:00.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:59 vm10 bash[19283]: cephadm 2026-03-07T10:09:58.252983+0000 mgr.a (mgr.14156) 71 : cephadm [INF] Deploying daemon osd.0 on vm02 2026-03-07T10:10:00.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:59 vm10 bash[19283]: cluster 2026-03-07T10:09:58.274680+0000 mgr.a (mgr.14156) 72 : cluster [DBG] pgmap v36: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:10:00.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:09:59 vm10 bash[19283]: cluster 2026-03-07T10:09:58.274680+0000 mgr.a (mgr.14156) 72 : cluster [DBG] pgmap v36: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:10:00.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:00 vm10 bash[19283]: audit 2026-03-07T10:09:59.783175+0000 mon.a (mon.0) 239 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:10:00.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:00 vm10 bash[19283]: audit 2026-03-07T10:09:59.783175+0000 mon.a (mon.0) 239 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:10:00.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:00 vm10 bash[19283]: audit 2026-03-07T10:09:59.788403+0000 mon.a (mon.0) 240 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:00.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:00 vm10 bash[19283]: audit 2026-03-07T10:09:59.788403+0000 mon.a (mon.0) 240 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:00.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:00 vm10 bash[19283]: audit 2026-03-07T10:09:59.794675+0000 mon.a (mon.0) 241 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:00.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:00 vm10 bash[19283]: audit 2026-03-07T10:09:59.794675+0000 mon.a (mon.0) 241 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:00.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:00 vm10 bash[19283]: cluster 2026-03-07T10:10:00.000095+0000 mon.a (mon.0) 242 : cluster [INF] overall HEALTH_OK 2026-03-07T10:10:00.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:00 vm10 bash[19283]: cluster 2026-03-07T10:10:00.000095+0000 mon.a (mon.0) 242 : cluster [INF] overall HEALTH_OK 2026-03-07T10:10:00.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:00 vm02 bash[17011]: audit 2026-03-07T10:09:59.783175+0000 mon.a (mon.0) 239 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:10:00.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:00 vm02 bash[17011]: audit 2026-03-07T10:09:59.783175+0000 mon.a (mon.0) 239 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:10:00.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:00 vm02 bash[17011]: audit 2026-03-07T10:09:59.788403+0000 mon.a (mon.0) 240 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:00.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:00 vm02 bash[17011]: audit 2026-03-07T10:09:59.788403+0000 mon.a (mon.0) 240 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:00.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:00 vm02 bash[17011]: audit 2026-03-07T10:09:59.794675+0000 mon.a (mon.0) 241 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:00.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:00 vm02 bash[17011]: audit 2026-03-07T10:09:59.794675+0000 mon.a (mon.0) 241 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:00.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:00 vm02 bash[17011]: cluster 2026-03-07T10:10:00.000095+0000 mon.a (mon.0) 242 : cluster [INF] overall HEALTH_OK 2026-03-07T10:10:00.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:00 vm02 bash[17011]: cluster 2026-03-07T10:10:00.000095+0000 mon.a (mon.0) 242 : cluster [INF] overall HEALTH_OK 2026-03-07T10:10:01.937 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:01 vm02 bash[17011]: cluster 2026-03-07T10:10:00.274841+0000 mgr.a (mgr.14156) 73 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:10:01.937 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:01 vm02 bash[17011]: cluster 2026-03-07T10:10:00.274841+0000 mgr.a (mgr.14156) 73 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:10:02.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:01 vm10 bash[19283]: cluster 2026-03-07T10:10:00.274841+0000 mgr.a (mgr.14156) 73 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:10:02.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:01 vm10 bash[19283]: cluster 2026-03-07T10:10:00.274841+0000 mgr.a (mgr.14156) 73 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:10:04.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:03 vm10 bash[19283]: cluster 2026-03-07T10:10:02.275035+0000 mgr.a (mgr.14156) 74 : cluster [DBG] pgmap v38: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:10:04.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:03 vm10 bash[19283]: cluster 2026-03-07T10:10:02.275035+0000 mgr.a (mgr.14156) 74 : cluster [DBG] pgmap v38: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:10:04.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:03 vm02 bash[17011]: cluster 2026-03-07T10:10:02.275035+0000 mgr.a (mgr.14156) 74 : cluster [DBG] pgmap v38: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:10:04.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:03 vm02 bash[17011]: cluster 2026-03-07T10:10:02.275035+0000 mgr.a (mgr.14156) 74 : cluster [DBG] pgmap v38: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:10:05.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:04 vm10 bash[19283]: audit 2026-03-07T10:10:03.863395+0000 mon.a (mon.0) 243 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3942075313,v1:192.168.123.102:6803/3942075313]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-07T10:10:05.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:04 vm10 bash[19283]: audit 2026-03-07T10:10:03.863395+0000 mon.a (mon.0) 243 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3942075313,v1:192.168.123.102:6803/3942075313]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-07T10:10:05.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:04 vm02 bash[17011]: audit 2026-03-07T10:10:03.863395+0000 mon.a (mon.0) 243 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3942075313,v1:192.168.123.102:6803/3942075313]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-07T10:10:05.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:04 vm02 bash[17011]: audit 2026-03-07T10:10:03.863395+0000 mon.a (mon.0) 243 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3942075313,v1:192.168.123.102:6803/3942075313]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-07T10:10:06.022 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:05 vm02 bash[17011]: cluster 2026-03-07T10:10:04.275217+0000 mgr.a (mgr.14156) 75 : cluster [DBG] pgmap v39: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:10:06.022 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:05 vm02 bash[17011]: cluster 2026-03-07T10:10:04.275217+0000 mgr.a (mgr.14156) 75 : cluster [DBG] pgmap v39: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:10:06.022 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:05 vm02 bash[17011]: audit 2026-03-07T10:10:04.859916+0000 mon.a (mon.0) 244 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3942075313,v1:192.168.123.102:6803/3942075313]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-07T10:10:06.022 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:05 vm02 bash[17011]: audit 2026-03-07T10:10:04.859916+0000 mon.a (mon.0) 244 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3942075313,v1:192.168.123.102:6803/3942075313]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-07T10:10:06.022 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:05 vm02 bash[17011]: cluster 2026-03-07T10:10:04.861940+0000 mon.a (mon.0) 245 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-07T10:10:06.022 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:05 vm02 bash[17011]: cluster 2026-03-07T10:10:04.861940+0000 mon.a (mon.0) 245 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-07T10:10:06.022 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:05 vm02 bash[17011]: audit 2026-03-07T10:10:04.862117+0000 mon.a (mon.0) 246 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3942075313,v1:192.168.123.102:6803/3942075313]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-07T10:10:06.022 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:05 vm02 bash[17011]: audit 2026-03-07T10:10:04.862117+0000 mon.a (mon.0) 246 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3942075313,v1:192.168.123.102:6803/3942075313]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-07T10:10:06.022 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:05 vm02 bash[17011]: audit 2026-03-07T10:10:04.862218+0000 mon.a (mon.0) 247 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T10:10:06.022 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:05 vm02 bash[17011]: audit 2026-03-07T10:10:04.862218+0000 mon.a (mon.0) 247 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T10:10:06.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:05 vm10 bash[19283]: cluster 2026-03-07T10:10:04.275217+0000 mgr.a (mgr.14156) 75 : cluster [DBG] pgmap v39: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:10:06.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:05 vm10 bash[19283]: cluster 2026-03-07T10:10:04.275217+0000 mgr.a (mgr.14156) 75 : cluster [DBG] pgmap v39: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:10:06.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:05 vm10 bash[19283]: audit 2026-03-07T10:10:04.859916+0000 mon.a (mon.0) 244 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3942075313,v1:192.168.123.102:6803/3942075313]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-07T10:10:06.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:05 vm10 bash[19283]: audit 2026-03-07T10:10:04.859916+0000 mon.a (mon.0) 244 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3942075313,v1:192.168.123.102:6803/3942075313]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-07T10:10:06.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:05 vm10 bash[19283]: cluster 2026-03-07T10:10:04.861940+0000 mon.a (mon.0) 245 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-07T10:10:06.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:05 vm10 bash[19283]: cluster 2026-03-07T10:10:04.861940+0000 mon.a (mon.0) 245 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-07T10:10:06.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:05 vm10 bash[19283]: audit 2026-03-07T10:10:04.862117+0000 mon.a (mon.0) 246 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3942075313,v1:192.168.123.102:6803/3942075313]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-07T10:10:06.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:05 vm10 bash[19283]: audit 2026-03-07T10:10:04.862117+0000 mon.a (mon.0) 246 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3942075313,v1:192.168.123.102:6803/3942075313]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-07T10:10:06.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:05 vm10 bash[19283]: audit 2026-03-07T10:10:04.862218+0000 mon.a (mon.0) 247 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T10:10:06.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:05 vm10 bash[19283]: audit 2026-03-07T10:10:04.862218+0000 mon.a (mon.0) 247 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T10:10:07.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:06 vm10 bash[19283]: cluster 2026-03-07T10:10:04.837814+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T10:10:07.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:06 vm10 bash[19283]: cluster 2026-03-07T10:10:04.837814+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T10:10:07.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:06 vm10 bash[19283]: cluster 2026-03-07T10:10:04.837868+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T10:10:07.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:06 vm10 bash[19283]: cluster 2026-03-07T10:10:04.837868+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T10:10:07.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:06 vm10 bash[19283]: audit 2026-03-07T10:10:05.865325+0000 mon.a (mon.0) 248 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3942075313,v1:192.168.123.102:6803/3942075313]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-07T10:10:07.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:06 vm10 bash[19283]: audit 2026-03-07T10:10:05.865325+0000 mon.a (mon.0) 248 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3942075313,v1:192.168.123.102:6803/3942075313]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-07T10:10:07.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:06 vm10 bash[19283]: cluster 2026-03-07T10:10:05.867480+0000 mon.a (mon.0) 249 : cluster [DBG] osdmap e7: 1 total, 0 up, 1 in 2026-03-07T10:10:07.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:06 vm10 bash[19283]: cluster 2026-03-07T10:10:05.867480+0000 mon.a (mon.0) 249 : cluster [DBG] osdmap e7: 1 total, 0 up, 1 in 2026-03-07T10:10:07.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:06 vm10 bash[19283]: audit 2026-03-07T10:10:05.870260+0000 mon.a (mon.0) 250 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T10:10:07.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:06 vm10 bash[19283]: audit 2026-03-07T10:10:05.870260+0000 mon.a (mon.0) 250 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T10:10:07.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:06 vm10 bash[19283]: audit 2026-03-07T10:10:05.876533+0000 mon.a (mon.0) 251 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T10:10:07.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:06 vm10 bash[19283]: audit 2026-03-07T10:10:05.876533+0000 mon.a (mon.0) 251 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T10:10:07.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:06 vm10 bash[19283]: audit 2026-03-07T10:10:05.924581+0000 mon.a (mon.0) 252 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:07.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:06 vm10 bash[19283]: audit 2026-03-07T10:10:05.924581+0000 mon.a (mon.0) 252 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:07.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:06 vm10 bash[19283]: audit 2026-03-07T10:10:05.929599+0000 mon.a (mon.0) 253 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:07.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:06 vm10 bash[19283]: audit 2026-03-07T10:10:05.929599+0000 mon.a (mon.0) 253 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:07.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:06 vm10 bash[19283]: audit 2026-03-07T10:10:06.260004+0000 mon.a (mon.0) 254 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:10:07.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:06 vm10 bash[19283]: audit 2026-03-07T10:10:06.260004+0000 mon.a (mon.0) 254 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:10:07.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:06 vm10 bash[19283]: audit 2026-03-07T10:10:06.260849+0000 mon.a (mon.0) 255 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:10:07.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:06 vm10 bash[19283]: audit 2026-03-07T10:10:06.260849+0000 mon.a (mon.0) 255 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:10:07.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:06 vm10 bash[19283]: audit 2026-03-07T10:10:06.267079+0000 mon.a (mon.0) 256 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:07.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:06 vm10 bash[19283]: audit 2026-03-07T10:10:06.267079+0000 mon.a (mon.0) 256 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:07.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:06 vm10 bash[19283]: audit 2026-03-07T10:10:06.872196+0000 mon.a (mon.0) 257 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T10:10:07.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:06 vm10 bash[19283]: audit 2026-03-07T10:10:06.872196+0000 mon.a (mon.0) 257 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T10:10:07.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:06 vm02 bash[17011]: cluster 2026-03-07T10:10:04.837814+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T10:10:07.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:06 vm02 bash[17011]: cluster 2026-03-07T10:10:04.837814+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T10:10:07.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:06 vm02 bash[17011]: cluster 2026-03-07T10:10:04.837868+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T10:10:07.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:06 vm02 bash[17011]: cluster 2026-03-07T10:10:04.837868+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T10:10:07.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:06 vm02 bash[17011]: audit 2026-03-07T10:10:05.865325+0000 mon.a (mon.0) 248 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3942075313,v1:192.168.123.102:6803/3942075313]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-07T10:10:07.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:06 vm02 bash[17011]: audit 2026-03-07T10:10:05.865325+0000 mon.a (mon.0) 248 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3942075313,v1:192.168.123.102:6803/3942075313]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-07T10:10:07.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:06 vm02 bash[17011]: cluster 2026-03-07T10:10:05.867480+0000 mon.a (mon.0) 249 : cluster [DBG] osdmap e7: 1 total, 0 up, 1 in 2026-03-07T10:10:07.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:06 vm02 bash[17011]: cluster 2026-03-07T10:10:05.867480+0000 mon.a (mon.0) 249 : cluster [DBG] osdmap e7: 1 total, 0 up, 1 in 2026-03-07T10:10:07.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:06 vm02 bash[17011]: audit 2026-03-07T10:10:05.870260+0000 mon.a (mon.0) 250 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T10:10:07.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:06 vm02 bash[17011]: audit 2026-03-07T10:10:05.870260+0000 mon.a (mon.0) 250 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T10:10:07.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:06 vm02 bash[17011]: audit 2026-03-07T10:10:05.876533+0000 mon.a (mon.0) 251 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T10:10:07.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:06 vm02 bash[17011]: audit 2026-03-07T10:10:05.876533+0000 mon.a (mon.0) 251 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T10:10:07.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:06 vm02 bash[17011]: audit 2026-03-07T10:10:05.924581+0000 mon.a (mon.0) 252 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:07.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:06 vm02 bash[17011]: audit 2026-03-07T10:10:05.924581+0000 mon.a (mon.0) 252 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:07.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:06 vm02 bash[17011]: audit 2026-03-07T10:10:05.929599+0000 mon.a (mon.0) 253 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:07.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:06 vm02 bash[17011]: audit 2026-03-07T10:10:05.929599+0000 mon.a (mon.0) 253 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:07.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:06 vm02 bash[17011]: audit 2026-03-07T10:10:06.260004+0000 mon.a (mon.0) 254 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:10:07.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:06 vm02 bash[17011]: audit 2026-03-07T10:10:06.260004+0000 mon.a (mon.0) 254 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:10:07.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:06 vm02 bash[17011]: audit 2026-03-07T10:10:06.260849+0000 mon.a (mon.0) 255 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:10:07.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:06 vm02 bash[17011]: audit 2026-03-07T10:10:06.260849+0000 mon.a (mon.0) 255 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:10:07.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:06 vm02 bash[17011]: audit 2026-03-07T10:10:06.267079+0000 mon.a (mon.0) 256 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:07.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:06 vm02 bash[17011]: audit 2026-03-07T10:10:06.267079+0000 mon.a (mon.0) 256 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:07.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:06 vm02 bash[17011]: audit 2026-03-07T10:10:06.872196+0000 mon.a (mon.0) 257 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T10:10:07.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:06 vm02 bash[17011]: audit 2026-03-07T10:10:06.872196+0000 mon.a (mon.0) 257 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T10:10:07.690 INFO:teuthology.orchestra.run.vm02.stdout:Created osd(s) 0 on host 'vm02' 2026-03-07T10:10:07.779 DEBUG:teuthology.orchestra.run.vm02:osd.0> sudo journalctl -f -n 0 -u ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@osd.0.service 2026-03-07T10:10:07.780 INFO:tasks.cephadm:Deploying osd.1 on vm10 with /dev/vde... 2026-03-07T10:10:07.780 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- lvm zap /dev/vde 2026-03-07T10:10:07.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:07 vm02 bash[17011]: cluster 2026-03-07T10:10:06.275409+0000 mgr.a (mgr.14156) 76 : cluster [DBG] pgmap v42: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:10:07.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:07 vm02 bash[17011]: cluster 2026-03-07T10:10:06.275409+0000 mgr.a (mgr.14156) 76 : cluster [DBG] pgmap v42: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:10:07.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:07 vm02 bash[17011]: cluster 2026-03-07T10:10:06.928215+0000 mon.a (mon.0) 258 : cluster [INF] osd.0 [v2:192.168.123.102:6802/3942075313,v1:192.168.123.102:6803/3942075313] boot 2026-03-07T10:10:07.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:07 vm02 bash[17011]: cluster 2026-03-07T10:10:06.928215+0000 mon.a (mon.0) 258 : cluster [INF] osd.0 [v2:192.168.123.102:6802/3942075313,v1:192.168.123.102:6803/3942075313] boot 2026-03-07T10:10:07.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:07 vm02 bash[17011]: cluster 2026-03-07T10:10:06.928316+0000 mon.a (mon.0) 259 : cluster [DBG] osdmap e8: 1 total, 1 up, 1 in 2026-03-07T10:10:07.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:07 vm02 bash[17011]: cluster 2026-03-07T10:10:06.928316+0000 mon.a (mon.0) 259 : cluster [DBG] osdmap e8: 1 total, 1 up, 1 in 2026-03-07T10:10:07.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:07 vm02 bash[17011]: audit 2026-03-07T10:10:06.928428+0000 mon.a (mon.0) 260 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T10:10:07.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:07 vm02 bash[17011]: audit 2026-03-07T10:10:06.928428+0000 mon.a (mon.0) 260 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T10:10:07.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:07 vm02 bash[17011]: audit 2026-03-07T10:10:07.672660+0000 mon.a (mon.0) 261 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:10:07.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:07 vm02 bash[17011]: audit 2026-03-07T10:10:07.672660+0000 mon.a (mon.0) 261 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:10:07.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:07 vm02 bash[17011]: audit 2026-03-07T10:10:07.677957+0000 mon.a (mon.0) 262 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:07.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:07 vm02 bash[17011]: audit 2026-03-07T10:10:07.677957+0000 mon.a (mon.0) 262 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:07.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:07 vm02 bash[17011]: audit 2026-03-07T10:10:07.682191+0000 mon.a (mon.0) 263 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:07.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:07 vm02 bash[17011]: audit 2026-03-07T10:10:07.682191+0000 mon.a (mon.0) 263 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:08.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:07 vm10 bash[19283]: cluster 2026-03-07T10:10:06.275409+0000 mgr.a (mgr.14156) 76 : cluster [DBG] pgmap v42: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:10:08.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:07 vm10 bash[19283]: cluster 2026-03-07T10:10:06.275409+0000 mgr.a (mgr.14156) 76 : cluster [DBG] pgmap v42: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T10:10:08.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:07 vm10 bash[19283]: cluster 2026-03-07T10:10:06.928215+0000 mon.a (mon.0) 258 : cluster [INF] osd.0 [v2:192.168.123.102:6802/3942075313,v1:192.168.123.102:6803/3942075313] boot 2026-03-07T10:10:08.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:07 vm10 bash[19283]: cluster 2026-03-07T10:10:06.928215+0000 mon.a (mon.0) 258 : cluster [INF] osd.0 [v2:192.168.123.102:6802/3942075313,v1:192.168.123.102:6803/3942075313] boot 2026-03-07T10:10:08.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:07 vm10 bash[19283]: cluster 2026-03-07T10:10:06.928316+0000 mon.a (mon.0) 259 : cluster [DBG] osdmap e8: 1 total, 1 up, 1 in 2026-03-07T10:10:08.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:07 vm10 bash[19283]: cluster 2026-03-07T10:10:06.928316+0000 mon.a (mon.0) 259 : cluster [DBG] osdmap e8: 1 total, 1 up, 1 in 2026-03-07T10:10:08.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:07 vm10 bash[19283]: audit 2026-03-07T10:10:06.928428+0000 mon.a (mon.0) 260 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T10:10:08.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:07 vm10 bash[19283]: audit 2026-03-07T10:10:06.928428+0000 mon.a (mon.0) 260 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T10:10:08.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:07 vm10 bash[19283]: audit 2026-03-07T10:10:07.672660+0000 mon.a (mon.0) 261 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:10:08.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:07 vm10 bash[19283]: audit 2026-03-07T10:10:07.672660+0000 mon.a (mon.0) 261 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:10:08.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:07 vm10 bash[19283]: audit 2026-03-07T10:10:07.677957+0000 mon.a (mon.0) 262 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:08.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:07 vm10 bash[19283]: audit 2026-03-07T10:10:07.677957+0000 mon.a (mon.0) 262 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:08.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:07 vm10 bash[19283]: audit 2026-03-07T10:10:07.682191+0000 mon.a (mon.0) 263 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:08.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:07 vm10 bash[19283]: audit 2026-03-07T10:10:07.682191+0000 mon.a (mon.0) 263 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:09.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:09 vm10 bash[19283]: cluster 2026-03-07T10:10:08.275677+0000 mgr.a (mgr.14156) 77 : cluster [DBG] pgmap v44: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:09.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:09 vm10 bash[19283]: cluster 2026-03-07T10:10:08.275677+0000 mgr.a (mgr.14156) 77 : cluster [DBG] pgmap v44: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:09.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:09 vm10 bash[19283]: cluster 2026-03-07T10:10:08.685349+0000 mon.a (mon.0) 264 : cluster [DBG] osdmap e9: 1 total, 1 up, 1 in 2026-03-07T10:10:09.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:09 vm10 bash[19283]: cluster 2026-03-07T10:10:08.685349+0000 mon.a (mon.0) 264 : cluster [DBG] osdmap e9: 1 total, 1 up, 1 in 2026-03-07T10:10:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:09 vm02 bash[17011]: cluster 2026-03-07T10:10:08.275677+0000 mgr.a (mgr.14156) 77 : cluster [DBG] pgmap v44: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:09 vm02 bash[17011]: cluster 2026-03-07T10:10:08.275677+0000 mgr.a (mgr.14156) 77 : cluster [DBG] pgmap v44: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:09 vm02 bash[17011]: cluster 2026-03-07T10:10:08.685349+0000 mon.a (mon.0) 264 : cluster [DBG] osdmap e9: 1 total, 1 up, 1 in 2026-03-07T10:10:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:09 vm02 bash[17011]: cluster 2026-03-07T10:10:08.685349+0000 mon.a (mon.0) 264 : cluster [DBG] osdmap e9: 1 total, 1 up, 1 in 2026-03-07T10:10:11.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:11 vm10 bash[19283]: cluster 2026-03-07T10:10:10.275982+0000 mgr.a (mgr.14156) 78 : cluster [DBG] pgmap v46: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:11.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:11 vm10 bash[19283]: cluster 2026-03-07T10:10:10.275982+0000 mgr.a (mgr.14156) 78 : cluster [DBG] pgmap v46: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:11.961 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:11 vm02 bash[17011]: cluster 2026-03-07T10:10:10.275982+0000 mgr.a (mgr.14156) 78 : cluster [DBG] pgmap v46: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:11.961 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:11 vm02 bash[17011]: cluster 2026-03-07T10:10:10.275982+0000 mgr.a (mgr.14156) 78 : cluster [DBG] pgmap v46: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:12.557 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.b/config 2026-03-07T10:10:13.752 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-07T10:10:13.771 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph orch daemon add osd vm10:/dev/vde 2026-03-07T10:10:13.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:13 vm10 bash[19283]: cluster 2026-03-07T10:10:12.276202+0000 mgr.a (mgr.14156) 79 : cluster [DBG] pgmap v47: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:13.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:13 vm10 bash[19283]: cluster 2026-03-07T10:10:12.276202+0000 mgr.a (mgr.14156) 79 : cluster [DBG] pgmap v47: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:13.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:13 vm02 bash[17011]: cluster 2026-03-07T10:10:12.276202+0000 mgr.a (mgr.14156) 79 : cluster [DBG] pgmap v47: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:13.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:13 vm02 bash[17011]: cluster 2026-03-07T10:10:12.276202+0000 mgr.a (mgr.14156) 79 : cluster [DBG] pgmap v47: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:15.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:14 vm10 bash[19283]: cephadm 2026-03-07T10:10:13.712262+0000 mgr.a (mgr.14156) 80 : cephadm [INF] Detected new or changed devices on vm02 2026-03-07T10:10:15.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:14 vm10 bash[19283]: cephadm 2026-03-07T10:10:13.712262+0000 mgr.a (mgr.14156) 80 : cephadm [INF] Detected new or changed devices on vm02 2026-03-07T10:10:15.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:14 vm10 bash[19283]: audit 2026-03-07T10:10:13.732177+0000 mon.a (mon.0) 265 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:15.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:14 vm10 bash[19283]: audit 2026-03-07T10:10:13.732177+0000 mon.a (mon.0) 265 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:15.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:14 vm10 bash[19283]: audit 2026-03-07T10:10:13.739150+0000 mon.a (mon.0) 266 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:15.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:14 vm10 bash[19283]: audit 2026-03-07T10:10:13.739150+0000 mon.a (mon.0) 266 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:15.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:14 vm10 bash[19283]: audit 2026-03-07T10:10:13.740027+0000 mon.a (mon.0) 267 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-07T10:10:15.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:14 vm10 bash[19283]: audit 2026-03-07T10:10:13.740027+0000 mon.a (mon.0) 267 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-07T10:10:15.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:14 vm10 bash[19283]: cephadm 2026-03-07T10:10:13.740495+0000 mgr.a (mgr.14156) 81 : cephadm [INF] Adjusting osd_memory_target on vm02 to 455.7M 2026-03-07T10:10:15.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:14 vm10 bash[19283]: cephadm 2026-03-07T10:10:13.740495+0000 mgr.a (mgr.14156) 81 : cephadm [INF] Adjusting osd_memory_target on vm02 to 455.7M 2026-03-07T10:10:15.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:14 vm10 bash[19283]: cephadm 2026-03-07T10:10:13.740990+0000 mgr.a (mgr.14156) 82 : cephadm [WRN] Unable to set osd_memory_target on vm02 to 477921689: error parsing value: Value '477921689' is below minimum 939524096 2026-03-07T10:10:15.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:14 vm10 bash[19283]: cephadm 2026-03-07T10:10:13.740990+0000 mgr.a (mgr.14156) 82 : cephadm [WRN] Unable to set osd_memory_target on vm02 to 477921689: error parsing value: Value '477921689' is below minimum 939524096 2026-03-07T10:10:15.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:14 vm10 bash[19283]: audit 2026-03-07T10:10:13.741362+0000 mon.a (mon.0) 268 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:10:15.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:14 vm10 bash[19283]: audit 2026-03-07T10:10:13.741362+0000 mon.a (mon.0) 268 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:10:15.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:14 vm10 bash[19283]: audit 2026-03-07T10:10:13.741858+0000 mon.a (mon.0) 269 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:10:15.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:14 vm10 bash[19283]: audit 2026-03-07T10:10:13.741858+0000 mon.a (mon.0) 269 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:10:15.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:14 vm10 bash[19283]: audit 2026-03-07T10:10:13.745567+0000 mon.a (mon.0) 270 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:15.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:14 vm10 bash[19283]: audit 2026-03-07T10:10:13.745567+0000 mon.a (mon.0) 270 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:15.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:14 vm02 bash[17011]: cephadm 2026-03-07T10:10:13.712262+0000 mgr.a (mgr.14156) 80 : cephadm [INF] Detected new or changed devices on vm02 2026-03-07T10:10:15.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:14 vm02 bash[17011]: cephadm 2026-03-07T10:10:13.712262+0000 mgr.a (mgr.14156) 80 : cephadm [INF] Detected new or changed devices on vm02 2026-03-07T10:10:15.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:14 vm02 bash[17011]: audit 2026-03-07T10:10:13.732177+0000 mon.a (mon.0) 265 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:15.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:14 vm02 bash[17011]: audit 2026-03-07T10:10:13.732177+0000 mon.a (mon.0) 265 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:15.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:14 vm02 bash[17011]: audit 2026-03-07T10:10:13.739150+0000 mon.a (mon.0) 266 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:15.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:14 vm02 bash[17011]: audit 2026-03-07T10:10:13.739150+0000 mon.a (mon.0) 266 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:15.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:14 vm02 bash[17011]: audit 2026-03-07T10:10:13.740027+0000 mon.a (mon.0) 267 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-07T10:10:15.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:14 vm02 bash[17011]: audit 2026-03-07T10:10:13.740027+0000 mon.a (mon.0) 267 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-07T10:10:15.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:14 vm02 bash[17011]: cephadm 2026-03-07T10:10:13.740495+0000 mgr.a (mgr.14156) 81 : cephadm [INF] Adjusting osd_memory_target on vm02 to 455.7M 2026-03-07T10:10:15.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:14 vm02 bash[17011]: cephadm 2026-03-07T10:10:13.740495+0000 mgr.a (mgr.14156) 81 : cephadm [INF] Adjusting osd_memory_target on vm02 to 455.7M 2026-03-07T10:10:15.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:14 vm02 bash[17011]: cephadm 2026-03-07T10:10:13.740990+0000 mgr.a (mgr.14156) 82 : cephadm [WRN] Unable to set osd_memory_target on vm02 to 477921689: error parsing value: Value '477921689' is below minimum 939524096 2026-03-07T10:10:15.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:14 vm02 bash[17011]: cephadm 2026-03-07T10:10:13.740990+0000 mgr.a (mgr.14156) 82 : cephadm [WRN] Unable to set osd_memory_target on vm02 to 477921689: error parsing value: Value '477921689' is below minimum 939524096 2026-03-07T10:10:15.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:14 vm02 bash[17011]: audit 2026-03-07T10:10:13.741362+0000 mon.a (mon.0) 268 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:10:15.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:14 vm02 bash[17011]: audit 2026-03-07T10:10:13.741362+0000 mon.a (mon.0) 268 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:10:15.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:14 vm02 bash[17011]: audit 2026-03-07T10:10:13.741858+0000 mon.a (mon.0) 269 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:10:15.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:14 vm02 bash[17011]: audit 2026-03-07T10:10:13.741858+0000 mon.a (mon.0) 269 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:10:15.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:14 vm02 bash[17011]: audit 2026-03-07T10:10:13.745567+0000 mon.a (mon.0) 270 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:15.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:14 vm02 bash[17011]: audit 2026-03-07T10:10:13.745567+0000 mon.a (mon.0) 270 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:16.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:15 vm10 bash[19283]: cluster 2026-03-07T10:10:14.276498+0000 mgr.a (mgr.14156) 83 : cluster [DBG] pgmap v48: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:16.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:15 vm10 bash[19283]: cluster 2026-03-07T10:10:14.276498+0000 mgr.a (mgr.14156) 83 : cluster [DBG] pgmap v48: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:16.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:15 vm02 bash[17011]: cluster 2026-03-07T10:10:14.276498+0000 mgr.a (mgr.14156) 83 : cluster [DBG] pgmap v48: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:16.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:15 vm02 bash[17011]: cluster 2026-03-07T10:10:14.276498+0000 mgr.a (mgr.14156) 83 : cluster [DBG] pgmap v48: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:18.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:17 vm10 bash[19283]: cluster 2026-03-07T10:10:16.276743+0000 mgr.a (mgr.14156) 84 : cluster [DBG] pgmap v49: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:18.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:17 vm10 bash[19283]: cluster 2026-03-07T10:10:16.276743+0000 mgr.a (mgr.14156) 84 : cluster [DBG] pgmap v49: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:18.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:17 vm02 bash[17011]: cluster 2026-03-07T10:10:16.276743+0000 mgr.a (mgr.14156) 84 : cluster [DBG] pgmap v49: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:18.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:17 vm02 bash[17011]: cluster 2026-03-07T10:10:16.276743+0000 mgr.a (mgr.14156) 84 : cluster [DBG] pgmap v49: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:18.535 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.b/config 2026-03-07T10:10:20.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:19 vm10 bash[19283]: cluster 2026-03-07T10:10:18.277057+0000 mgr.a (mgr.14156) 85 : cluster [DBG] pgmap v50: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:20.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:19 vm10 bash[19283]: cluster 2026-03-07T10:10:18.277057+0000 mgr.a (mgr.14156) 85 : cluster [DBG] pgmap v50: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:20.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:19 vm10 bash[19283]: audit 2026-03-07T10:10:18.901959+0000 mgr.a (mgr.14156) 86 : audit [DBG] from='client.24127 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm10:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:10:20.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:19 vm10 bash[19283]: audit 2026-03-07T10:10:18.901959+0000 mgr.a (mgr.14156) 86 : audit [DBG] from='client.24127 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm10:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:10:20.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:19 vm10 bash[19283]: audit 2026-03-07T10:10:18.903365+0000 mon.a (mon.0) 271 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T10:10:20.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:19 vm10 bash[19283]: audit 2026-03-07T10:10:18.903365+0000 mon.a (mon.0) 271 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T10:10:20.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:19 vm10 bash[19283]: audit 2026-03-07T10:10:18.904527+0000 mon.a (mon.0) 272 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-07T10:10:20.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:19 vm10 bash[19283]: audit 2026-03-07T10:10:18.904527+0000 mon.a (mon.0) 272 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-07T10:10:20.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:19 vm10 bash[19283]: audit 2026-03-07T10:10:18.904907+0000 mon.a (mon.0) 273 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:10:20.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:19 vm10 bash[19283]: audit 2026-03-07T10:10:18.904907+0000 mon.a (mon.0) 273 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:10:20.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:19 vm02 bash[17011]: cluster 2026-03-07T10:10:18.277057+0000 mgr.a (mgr.14156) 85 : cluster [DBG] pgmap v50: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:20.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:19 vm02 bash[17011]: cluster 2026-03-07T10:10:18.277057+0000 mgr.a (mgr.14156) 85 : cluster [DBG] pgmap v50: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:20.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:19 vm02 bash[17011]: audit 2026-03-07T10:10:18.901959+0000 mgr.a (mgr.14156) 86 : audit [DBG] from='client.24127 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm10:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:10:20.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:19 vm02 bash[17011]: audit 2026-03-07T10:10:18.901959+0000 mgr.a (mgr.14156) 86 : audit [DBG] from='client.24127 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm10:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:10:20.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:19 vm02 bash[17011]: audit 2026-03-07T10:10:18.903365+0000 mon.a (mon.0) 271 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T10:10:20.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:19 vm02 bash[17011]: audit 2026-03-07T10:10:18.903365+0000 mon.a (mon.0) 271 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T10:10:20.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:19 vm02 bash[17011]: audit 2026-03-07T10:10:18.904527+0000 mon.a (mon.0) 272 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-07T10:10:20.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:19 vm02 bash[17011]: audit 2026-03-07T10:10:18.904527+0000 mon.a (mon.0) 272 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-07T10:10:20.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:19 vm02 bash[17011]: audit 2026-03-07T10:10:18.904907+0000 mon.a (mon.0) 273 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:10:20.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:19 vm02 bash[17011]: audit 2026-03-07T10:10:18.904907+0000 mon.a (mon.0) 273 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:10:22.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:21 vm10 bash[19283]: cluster 2026-03-07T10:10:20.277296+0000 mgr.a (mgr.14156) 87 : cluster [DBG] pgmap v51: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:22.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:21 vm10 bash[19283]: cluster 2026-03-07T10:10:20.277296+0000 mgr.a (mgr.14156) 87 : cluster [DBG] pgmap v51: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:22.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:21 vm02 bash[17011]: cluster 2026-03-07T10:10:20.277296+0000 mgr.a (mgr.14156) 87 : cluster [DBG] pgmap v51: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:22.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:21 vm02 bash[17011]: cluster 2026-03-07T10:10:20.277296+0000 mgr.a (mgr.14156) 87 : cluster [DBG] pgmap v51: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:24.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:23 vm10 bash[19283]: cluster 2026-03-07T10:10:22.277510+0000 mgr.a (mgr.14156) 88 : cluster [DBG] pgmap v52: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:24.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:23 vm10 bash[19283]: cluster 2026-03-07T10:10:22.277510+0000 mgr.a (mgr.14156) 88 : cluster [DBG] pgmap v52: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:24.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:23 vm02 bash[17011]: cluster 2026-03-07T10:10:22.277510+0000 mgr.a (mgr.14156) 88 : cluster [DBG] pgmap v52: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:24.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:23 vm02 bash[17011]: cluster 2026-03-07T10:10:22.277510+0000 mgr.a (mgr.14156) 88 : cluster [DBG] pgmap v52: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:24.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:24 vm10 bash[19283]: audit 2026-03-07T10:10:24.577776+0000 mon.a (mon.0) 274 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4d550476-452d-4382-8bdf-ae95a10d4527"}]: dispatch 2026-03-07T10:10:24.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:24 vm10 bash[19283]: audit 2026-03-07T10:10:24.577776+0000 mon.a (mon.0) 274 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4d550476-452d-4382-8bdf-ae95a10d4527"}]: dispatch 2026-03-07T10:10:24.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:24 vm10 bash[19283]: audit 2026-03-07T10:10:24.578457+0000 mon.b (mon.1) 6 : audit [INF] from='client.? 192.168.123.110:0/3472466638' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4d550476-452d-4382-8bdf-ae95a10d4527"}]: dispatch 2026-03-07T10:10:24.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:24 vm10 bash[19283]: audit 2026-03-07T10:10:24.578457+0000 mon.b (mon.1) 6 : audit [INF] from='client.? 192.168.123.110:0/3472466638' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4d550476-452d-4382-8bdf-ae95a10d4527"}]: dispatch 2026-03-07T10:10:24.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:24 vm10 bash[19283]: audit 2026-03-07T10:10:24.581295+0000 mon.a (mon.0) 275 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "4d550476-452d-4382-8bdf-ae95a10d4527"}]': finished 2026-03-07T10:10:24.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:24 vm10 bash[19283]: audit 2026-03-07T10:10:24.581295+0000 mon.a (mon.0) 275 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "4d550476-452d-4382-8bdf-ae95a10d4527"}]': finished 2026-03-07T10:10:24.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:24 vm10 bash[19283]: cluster 2026-03-07T10:10:24.583701+0000 mon.a (mon.0) 276 : cluster [DBG] osdmap e10: 2 total, 1 up, 2 in 2026-03-07T10:10:24.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:24 vm10 bash[19283]: cluster 2026-03-07T10:10:24.583701+0000 mon.a (mon.0) 276 : cluster [DBG] osdmap e10: 2 total, 1 up, 2 in 2026-03-07T10:10:24.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:24 vm10 bash[19283]: audit 2026-03-07T10:10:24.583774+0000 mon.a (mon.0) 277 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T10:10:24.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:24 vm10 bash[19283]: audit 2026-03-07T10:10:24.583774+0000 mon.a (mon.0) 277 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T10:10:25.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:24 vm02 bash[17011]: audit 2026-03-07T10:10:24.577776+0000 mon.a (mon.0) 274 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4d550476-452d-4382-8bdf-ae95a10d4527"}]: dispatch 2026-03-07T10:10:25.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:24 vm02 bash[17011]: audit 2026-03-07T10:10:24.577776+0000 mon.a (mon.0) 274 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4d550476-452d-4382-8bdf-ae95a10d4527"}]: dispatch 2026-03-07T10:10:25.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:24 vm02 bash[17011]: audit 2026-03-07T10:10:24.578457+0000 mon.b (mon.1) 6 : audit [INF] from='client.? 192.168.123.110:0/3472466638' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4d550476-452d-4382-8bdf-ae95a10d4527"}]: dispatch 2026-03-07T10:10:25.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:24 vm02 bash[17011]: audit 2026-03-07T10:10:24.578457+0000 mon.b (mon.1) 6 : audit [INF] from='client.? 192.168.123.110:0/3472466638' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4d550476-452d-4382-8bdf-ae95a10d4527"}]: dispatch 2026-03-07T10:10:25.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:24 vm02 bash[17011]: audit 2026-03-07T10:10:24.581295+0000 mon.a (mon.0) 275 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "4d550476-452d-4382-8bdf-ae95a10d4527"}]': finished 2026-03-07T10:10:25.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:24 vm02 bash[17011]: audit 2026-03-07T10:10:24.581295+0000 mon.a (mon.0) 275 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "4d550476-452d-4382-8bdf-ae95a10d4527"}]': finished 2026-03-07T10:10:25.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:24 vm02 bash[17011]: cluster 2026-03-07T10:10:24.583701+0000 mon.a (mon.0) 276 : cluster [DBG] osdmap e10: 2 total, 1 up, 2 in 2026-03-07T10:10:25.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:24 vm02 bash[17011]: cluster 2026-03-07T10:10:24.583701+0000 mon.a (mon.0) 276 : cluster [DBG] osdmap e10: 2 total, 1 up, 2 in 2026-03-07T10:10:25.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:24 vm02 bash[17011]: audit 2026-03-07T10:10:24.583774+0000 mon.a (mon.0) 277 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T10:10:25.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:24 vm02 bash[17011]: audit 2026-03-07T10:10:24.583774+0000 mon.a (mon.0) 277 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T10:10:26.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:25 vm10 bash[19283]: cluster 2026-03-07T10:10:24.277744+0000 mgr.a (mgr.14156) 89 : cluster [DBG] pgmap v53: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:26.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:25 vm10 bash[19283]: cluster 2026-03-07T10:10:24.277744+0000 mgr.a (mgr.14156) 89 : cluster [DBG] pgmap v53: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:26.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:25 vm10 bash[19283]: audit 2026-03-07T10:10:25.275932+0000 mon.b (mon.1) 7 : audit [DBG] from='client.? 192.168.123.110:0/1034057114' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T10:10:26.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:25 vm10 bash[19283]: audit 2026-03-07T10:10:25.275932+0000 mon.b (mon.1) 7 : audit [DBG] from='client.? 192.168.123.110:0/1034057114' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T10:10:26.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:25 vm02 bash[17011]: cluster 2026-03-07T10:10:24.277744+0000 mgr.a (mgr.14156) 89 : cluster [DBG] pgmap v53: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:26.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:25 vm02 bash[17011]: cluster 2026-03-07T10:10:24.277744+0000 mgr.a (mgr.14156) 89 : cluster [DBG] pgmap v53: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:26.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:25 vm02 bash[17011]: audit 2026-03-07T10:10:25.275932+0000 mon.b (mon.1) 7 : audit [DBG] from='client.? 192.168.123.110:0/1034057114' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T10:10:26.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:25 vm02 bash[17011]: audit 2026-03-07T10:10:25.275932+0000 mon.b (mon.1) 7 : audit [DBG] from='client.? 192.168.123.110:0/1034057114' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T10:10:28.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:27 vm10 bash[19283]: cluster 2026-03-07T10:10:26.278002+0000 mgr.a (mgr.14156) 90 : cluster [DBG] pgmap v55: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:28.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:27 vm10 bash[19283]: cluster 2026-03-07T10:10:26.278002+0000 mgr.a (mgr.14156) 90 : cluster [DBG] pgmap v55: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:28.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:27 vm02 bash[17011]: cluster 2026-03-07T10:10:26.278002+0000 mgr.a (mgr.14156) 90 : cluster [DBG] pgmap v55: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:28.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:27 vm02 bash[17011]: cluster 2026-03-07T10:10:26.278002+0000 mgr.a (mgr.14156) 90 : cluster [DBG] pgmap v55: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:30.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:29 vm10 bash[19283]: cluster 2026-03-07T10:10:28.278307+0000 mgr.a (mgr.14156) 91 : cluster [DBG] pgmap v56: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:30.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:29 vm10 bash[19283]: cluster 2026-03-07T10:10:28.278307+0000 mgr.a (mgr.14156) 91 : cluster [DBG] pgmap v56: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:30.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:29 vm02 bash[17011]: cluster 2026-03-07T10:10:28.278307+0000 mgr.a (mgr.14156) 91 : cluster [DBG] pgmap v56: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:30.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:29 vm02 bash[17011]: cluster 2026-03-07T10:10:28.278307+0000 mgr.a (mgr.14156) 91 : cluster [DBG] pgmap v56: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:32.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:31 vm10 bash[19283]: cluster 2026-03-07T10:10:30.278591+0000 mgr.a (mgr.14156) 92 : cluster [DBG] pgmap v57: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:32.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:31 vm10 bash[19283]: cluster 2026-03-07T10:10:30.278591+0000 mgr.a (mgr.14156) 92 : cluster [DBG] pgmap v57: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:32.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:31 vm02 bash[17011]: cluster 2026-03-07T10:10:30.278591+0000 mgr.a (mgr.14156) 92 : cluster [DBG] pgmap v57: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:32.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:31 vm02 bash[17011]: cluster 2026-03-07T10:10:30.278591+0000 mgr.a (mgr.14156) 92 : cluster [DBG] pgmap v57: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:34.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:33 vm10 bash[19283]: cluster 2026-03-07T10:10:32.278785+0000 mgr.a (mgr.14156) 93 : cluster [DBG] pgmap v58: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:34.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:33 vm10 bash[19283]: cluster 2026-03-07T10:10:32.278785+0000 mgr.a (mgr.14156) 93 : cluster [DBG] pgmap v58: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:34.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:33 vm02 bash[17011]: cluster 2026-03-07T10:10:32.278785+0000 mgr.a (mgr.14156) 93 : cluster [DBG] pgmap v58: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:34.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:33 vm02 bash[17011]: cluster 2026-03-07T10:10:32.278785+0000 mgr.a (mgr.14156) 93 : cluster [DBG] pgmap v58: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:35.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:34 vm10 bash[19283]: audit 2026-03-07T10:10:34.523928+0000 mon.a (mon.0) 278 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-07T10:10:35.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:34 vm10 bash[19283]: audit 2026-03-07T10:10:34.523928+0000 mon.a (mon.0) 278 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-07T10:10:35.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:34 vm10 bash[19283]: audit 2026-03-07T10:10:34.524391+0000 mon.a (mon.0) 279 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:10:35.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:34 vm10 bash[19283]: audit 2026-03-07T10:10:34.524391+0000 mon.a (mon.0) 279 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:10:35.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:34 vm02 bash[17011]: audit 2026-03-07T10:10:34.523928+0000 mon.a (mon.0) 278 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-07T10:10:35.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:34 vm02 bash[17011]: audit 2026-03-07T10:10:34.523928+0000 mon.a (mon.0) 278 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-07T10:10:35.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:34 vm02 bash[17011]: audit 2026-03-07T10:10:34.524391+0000 mon.a (mon.0) 279 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:10:35.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:34 vm02 bash[17011]: audit 2026-03-07T10:10:34.524391+0000 mon.a (mon.0) 279 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:10:35.641 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:35 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:10:35.641 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:10:35 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:10:35.906 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:10:35 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:10:35.906 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:35 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:10:35.906 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:35 vm10 bash[19283]: cluster 2026-03-07T10:10:34.279008+0000 mgr.a (mgr.14156) 94 : cluster [DBG] pgmap v59: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:35.906 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:35 vm10 bash[19283]: cluster 2026-03-07T10:10:34.279008+0000 mgr.a (mgr.14156) 94 : cluster [DBG] pgmap v59: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:35.907 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:35 vm10 bash[19283]: cephadm 2026-03-07T10:10:34.524729+0000 mgr.a (mgr.14156) 95 : cephadm [INF] Deploying daemon osd.1 on vm10 2026-03-07T10:10:35.907 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:35 vm10 bash[19283]: cephadm 2026-03-07T10:10:34.524729+0000 mgr.a (mgr.14156) 95 : cephadm [INF] Deploying daemon osd.1 on vm10 2026-03-07T10:10:35.907 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:35 vm10 bash[19283]: audit 2026-03-07T10:10:35.843492+0000 mon.a (mon.0) 280 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:10:35.907 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:35 vm10 bash[19283]: audit 2026-03-07T10:10:35.843492+0000 mon.a (mon.0) 280 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:10:36.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:35 vm02 bash[17011]: cluster 2026-03-07T10:10:34.279008+0000 mgr.a (mgr.14156) 94 : cluster [DBG] pgmap v59: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:36.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:35 vm02 bash[17011]: cluster 2026-03-07T10:10:34.279008+0000 mgr.a (mgr.14156) 94 : cluster [DBG] pgmap v59: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:36.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:35 vm02 bash[17011]: cephadm 2026-03-07T10:10:34.524729+0000 mgr.a (mgr.14156) 95 : cephadm [INF] Deploying daemon osd.1 on vm10 2026-03-07T10:10:36.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:35 vm02 bash[17011]: cephadm 2026-03-07T10:10:34.524729+0000 mgr.a (mgr.14156) 95 : cephadm [INF] Deploying daemon osd.1 on vm10 2026-03-07T10:10:36.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:35 vm02 bash[17011]: audit 2026-03-07T10:10:35.843492+0000 mon.a (mon.0) 280 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:10:36.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:35 vm02 bash[17011]: audit 2026-03-07T10:10:35.843492+0000 mon.a (mon.0) 280 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:10:37.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:36 vm10 bash[19283]: audit 2026-03-07T10:10:35.848693+0000 mon.a (mon.0) 281 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:37.197 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:36 vm10 bash[19283]: audit 2026-03-07T10:10:35.848693+0000 mon.a (mon.0) 281 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:37.197 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:36 vm10 bash[19283]: audit 2026-03-07T10:10:35.866543+0000 mon.a (mon.0) 282 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:37.197 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:36 vm10 bash[19283]: audit 2026-03-07T10:10:35.866543+0000 mon.a (mon.0) 282 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:37.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:36 vm02 bash[17011]: audit 2026-03-07T10:10:35.848693+0000 mon.a (mon.0) 281 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:36 vm02 bash[17011]: audit 2026-03-07T10:10:35.848693+0000 mon.a (mon.0) 281 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:36 vm02 bash[17011]: audit 2026-03-07T10:10:35.866543+0000 mon.a (mon.0) 282 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:37.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:36 vm02 bash[17011]: audit 2026-03-07T10:10:35.866543+0000 mon.a (mon.0) 282 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:38.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:37 vm10 bash[19283]: cluster 2026-03-07T10:10:36.279224+0000 mgr.a (mgr.14156) 96 : cluster [DBG] pgmap v60: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:38.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:37 vm10 bash[19283]: cluster 2026-03-07T10:10:36.279224+0000 mgr.a (mgr.14156) 96 : cluster [DBG] pgmap v60: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:38.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:37 vm02 bash[17011]: cluster 2026-03-07T10:10:36.279224+0000 mgr.a (mgr.14156) 96 : cluster [DBG] pgmap v60: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:38.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:37 vm02 bash[17011]: cluster 2026-03-07T10:10:36.279224+0000 mgr.a (mgr.14156) 96 : cluster [DBG] pgmap v60: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:40.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:39 vm10 bash[19283]: cluster 2026-03-07T10:10:38.279479+0000 mgr.a (mgr.14156) 97 : cluster [DBG] pgmap v61: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:40.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:39 vm10 bash[19283]: cluster 2026-03-07T10:10:38.279479+0000 mgr.a (mgr.14156) 97 : cluster [DBG] pgmap v61: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:40.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:39 vm10 bash[19283]: audit 2026-03-07T10:10:39.824568+0000 mon.a (mon.0) 283 : audit [INF] from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-07T10:10:40.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:39 vm10 bash[19283]: audit 2026-03-07T10:10:39.824568+0000 mon.a (mon.0) 283 : audit [INF] from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-07T10:10:40.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:39 vm10 bash[19283]: audit 2026-03-07T10:10:39.825472+0000 mon.b (mon.1) 8 : audit [INF] from='osd.1 [v2:192.168.123.110:6800/1483542873,v1:192.168.123.110:6801/1483542873]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-07T10:10:40.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:39 vm10 bash[19283]: audit 2026-03-07T10:10:39.825472+0000 mon.b (mon.1) 8 : audit [INF] from='osd.1 [v2:192.168.123.110:6800/1483542873,v1:192.168.123.110:6801/1483542873]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-07T10:10:40.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:39 vm02 bash[17011]: cluster 2026-03-07T10:10:38.279479+0000 mgr.a (mgr.14156) 97 : cluster [DBG] pgmap v61: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:40.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:39 vm02 bash[17011]: cluster 2026-03-07T10:10:38.279479+0000 mgr.a (mgr.14156) 97 : cluster [DBG] pgmap v61: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:40.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:39 vm02 bash[17011]: audit 2026-03-07T10:10:39.824568+0000 mon.a (mon.0) 283 : audit [INF] from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-07T10:10:40.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:39 vm02 bash[17011]: audit 2026-03-07T10:10:39.824568+0000 mon.a (mon.0) 283 : audit [INF] from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-07T10:10:40.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:39 vm02 bash[17011]: audit 2026-03-07T10:10:39.825472+0000 mon.b (mon.1) 8 : audit [INF] from='osd.1 [v2:192.168.123.110:6800/1483542873,v1:192.168.123.110:6801/1483542873]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-07T10:10:40.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:39 vm02 bash[17011]: audit 2026-03-07T10:10:39.825472+0000 mon.b (mon.1) 8 : audit [INF] from='osd.1 [v2:192.168.123.110:6800/1483542873,v1:192.168.123.110:6801/1483542873]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-07T10:10:41.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:40 vm10 bash[19283]: audit 2026-03-07T10:10:39.875000+0000 mon.a (mon.0) 284 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-07T10:10:41.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:40 vm10 bash[19283]: audit 2026-03-07T10:10:39.875000+0000 mon.a (mon.0) 284 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-07T10:10:41.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:40 vm10 bash[19283]: cluster 2026-03-07T10:10:39.878000+0000 mon.a (mon.0) 285 : cluster [DBG] osdmap e11: 2 total, 1 up, 2 in 2026-03-07T10:10:41.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:40 vm10 bash[19283]: cluster 2026-03-07T10:10:39.878000+0000 mon.a (mon.0) 285 : cluster [DBG] osdmap e11: 2 total, 1 up, 2 in 2026-03-07T10:10:41.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:40 vm10 bash[19283]: audit 2026-03-07T10:10:39.878794+0000 mon.a (mon.0) 286 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T10:10:41.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:40 vm10 bash[19283]: audit 2026-03-07T10:10:39.878794+0000 mon.a (mon.0) 286 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T10:10:41.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:40 vm10 bash[19283]: audit 2026-03-07T10:10:39.878869+0000 mon.a (mon.0) 287 : audit [INF] from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-07T10:10:41.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:40 vm10 bash[19283]: audit 2026-03-07T10:10:39.878869+0000 mon.a (mon.0) 287 : audit [INF] from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-07T10:10:41.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:40 vm10 bash[19283]: audit 2026-03-07T10:10:39.879296+0000 mon.b (mon.1) 9 : audit [INF] from='osd.1 [v2:192.168.123.110:6800/1483542873,v1:192.168.123.110:6801/1483542873]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-07T10:10:41.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:40 vm10 bash[19283]: audit 2026-03-07T10:10:39.879296+0000 mon.b (mon.1) 9 : audit [INF] from='osd.1 [v2:192.168.123.110:6800/1483542873,v1:192.168.123.110:6801/1483542873]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-07T10:10:41.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:40 vm10 bash[19283]: audit 2026-03-07T10:10:40.878194+0000 mon.a (mon.0) 288 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm10", "root=default"]}]': finished 2026-03-07T10:10:41.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:40 vm10 bash[19283]: audit 2026-03-07T10:10:40.878194+0000 mon.a (mon.0) 288 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm10", "root=default"]}]': finished 2026-03-07T10:10:41.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:40 vm10 bash[19283]: cluster 2026-03-07T10:10:40.880701+0000 mon.a (mon.0) 289 : cluster [DBG] osdmap e12: 2 total, 1 up, 2 in 2026-03-07T10:10:41.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:40 vm10 bash[19283]: cluster 2026-03-07T10:10:40.880701+0000 mon.a (mon.0) 289 : cluster [DBG] osdmap e12: 2 total, 1 up, 2 in 2026-03-07T10:10:41.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:40 vm02 bash[17011]: audit 2026-03-07T10:10:39.875000+0000 mon.a (mon.0) 284 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-07T10:10:41.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:40 vm02 bash[17011]: audit 2026-03-07T10:10:39.875000+0000 mon.a (mon.0) 284 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-07T10:10:41.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:40 vm02 bash[17011]: cluster 2026-03-07T10:10:39.878000+0000 mon.a (mon.0) 285 : cluster [DBG] osdmap e11: 2 total, 1 up, 2 in 2026-03-07T10:10:41.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:40 vm02 bash[17011]: cluster 2026-03-07T10:10:39.878000+0000 mon.a (mon.0) 285 : cluster [DBG] osdmap e11: 2 total, 1 up, 2 in 2026-03-07T10:10:41.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:40 vm02 bash[17011]: audit 2026-03-07T10:10:39.878794+0000 mon.a (mon.0) 286 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T10:10:41.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:40 vm02 bash[17011]: audit 2026-03-07T10:10:39.878794+0000 mon.a (mon.0) 286 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T10:10:41.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:40 vm02 bash[17011]: audit 2026-03-07T10:10:39.878869+0000 mon.a (mon.0) 287 : audit [INF] from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-07T10:10:41.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:40 vm02 bash[17011]: audit 2026-03-07T10:10:39.878869+0000 mon.a (mon.0) 287 : audit [INF] from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-07T10:10:41.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:40 vm02 bash[17011]: audit 2026-03-07T10:10:39.879296+0000 mon.b (mon.1) 9 : audit [INF] from='osd.1 [v2:192.168.123.110:6800/1483542873,v1:192.168.123.110:6801/1483542873]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-07T10:10:41.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:40 vm02 bash[17011]: audit 2026-03-07T10:10:39.879296+0000 mon.b (mon.1) 9 : audit [INF] from='osd.1 [v2:192.168.123.110:6800/1483542873,v1:192.168.123.110:6801/1483542873]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-07T10:10:41.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:40 vm02 bash[17011]: audit 2026-03-07T10:10:40.878194+0000 mon.a (mon.0) 288 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm10", "root=default"]}]': finished 2026-03-07T10:10:41.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:40 vm02 bash[17011]: audit 2026-03-07T10:10:40.878194+0000 mon.a (mon.0) 288 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm10", "root=default"]}]': finished 2026-03-07T10:10:41.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:40 vm02 bash[17011]: cluster 2026-03-07T10:10:40.880701+0000 mon.a (mon.0) 289 : cluster [DBG] osdmap e12: 2 total, 1 up, 2 in 2026-03-07T10:10:41.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:40 vm02 bash[17011]: cluster 2026-03-07T10:10:40.880701+0000 mon.a (mon.0) 289 : cluster [DBG] osdmap e12: 2 total, 1 up, 2 in 2026-03-07T10:10:42.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:41 vm10 bash[19283]: cluster 2026-03-07T10:10:40.279847+0000 mgr.a (mgr.14156) 98 : cluster [DBG] pgmap v63: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:42.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:41 vm10 bash[19283]: cluster 2026-03-07T10:10:40.279847+0000 mgr.a (mgr.14156) 98 : cluster [DBG] pgmap v63: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:42.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:41 vm10 bash[19283]: audit 2026-03-07T10:10:40.880799+0000 mon.a (mon.0) 290 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T10:10:42.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:41 vm10 bash[19283]: audit 2026-03-07T10:10:40.880799+0000 mon.a (mon.0) 290 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T10:10:42.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:41 vm10 bash[19283]: audit 2026-03-07T10:10:40.897871+0000 mon.a (mon.0) 291 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T10:10:42.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:41 vm10 bash[19283]: audit 2026-03-07T10:10:40.897871+0000 mon.a (mon.0) 291 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T10:10:42.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:41 vm02 bash[17011]: cluster 2026-03-07T10:10:40.279847+0000 mgr.a (mgr.14156) 98 : cluster [DBG] pgmap v63: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:42.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:41 vm02 bash[17011]: cluster 2026-03-07T10:10:40.279847+0000 mgr.a (mgr.14156) 98 : cluster [DBG] pgmap v63: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:42.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:41 vm02 bash[17011]: audit 2026-03-07T10:10:40.880799+0000 mon.a (mon.0) 290 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T10:10:42.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:41 vm02 bash[17011]: audit 2026-03-07T10:10:40.880799+0000 mon.a (mon.0) 290 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T10:10:42.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:41 vm02 bash[17011]: audit 2026-03-07T10:10:40.897871+0000 mon.a (mon.0) 291 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T10:10:42.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:41 vm02 bash[17011]: audit 2026-03-07T10:10:40.897871+0000 mon.a (mon.0) 291 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T10:10:43.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:42 vm10 bash[19283]: cluster 2026-03-07T10:10:40.802008+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T10:10:43.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:42 vm10 bash[19283]: cluster 2026-03-07T10:10:40.802008+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T10:10:43.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:42 vm10 bash[19283]: cluster 2026-03-07T10:10:40.802064+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T10:10:43.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:42 vm10 bash[19283]: cluster 2026-03-07T10:10:40.802064+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T10:10:43.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:42 vm10 bash[19283]: audit 2026-03-07T10:10:41.892222+0000 mon.a (mon.0) 292 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T10:10:43.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:42 vm10 bash[19283]: audit 2026-03-07T10:10:41.892222+0000 mon.a (mon.0) 292 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T10:10:43.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:42 vm10 bash[19283]: cluster 2026-03-07T10:10:41.905629+0000 mon.a (mon.0) 293 : cluster [INF] osd.1 [v2:192.168.123.110:6800/1483542873,v1:192.168.123.110:6801/1483542873] boot 2026-03-07T10:10:43.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:42 vm10 bash[19283]: cluster 2026-03-07T10:10:41.905629+0000 mon.a (mon.0) 293 : cluster [INF] osd.1 [v2:192.168.123.110:6800/1483542873,v1:192.168.123.110:6801/1483542873] boot 2026-03-07T10:10:43.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:42 vm10 bash[19283]: cluster 2026-03-07T10:10:41.905656+0000 mon.a (mon.0) 294 : cluster [DBG] osdmap e13: 2 total, 2 up, 2 in 2026-03-07T10:10:43.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:42 vm10 bash[19283]: cluster 2026-03-07T10:10:41.905656+0000 mon.a (mon.0) 294 : cluster [DBG] osdmap e13: 2 total, 2 up, 2 in 2026-03-07T10:10:43.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:42 vm10 bash[19283]: audit 2026-03-07T10:10:41.905701+0000 mon.a (mon.0) 295 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T10:10:43.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:42 vm10 bash[19283]: audit 2026-03-07T10:10:41.905701+0000 mon.a (mon.0) 295 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T10:10:43.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:42 vm10 bash[19283]: audit 2026-03-07T10:10:42.074598+0000 mon.a (mon.0) 296 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:43.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:42 vm10 bash[19283]: audit 2026-03-07T10:10:42.074598+0000 mon.a (mon.0) 296 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:43.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:42 vm10 bash[19283]: audit 2026-03-07T10:10:42.077780+0000 mon.a (mon.0) 297 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:43.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:42 vm10 bash[19283]: audit 2026-03-07T10:10:42.077780+0000 mon.a (mon.0) 297 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:43.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:42 vm10 bash[19283]: audit 2026-03-07T10:10:42.078310+0000 mon.a (mon.0) 298 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:10:43.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:42 vm10 bash[19283]: audit 2026-03-07T10:10:42.078310+0000 mon.a (mon.0) 298 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:10:43.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:42 vm10 bash[19283]: audit 2026-03-07T10:10:42.078692+0000 mon.a (mon.0) 299 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:10:43.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:42 vm10 bash[19283]: audit 2026-03-07T10:10:42.078692+0000 mon.a (mon.0) 299 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:10:43.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:42 vm10 bash[19283]: audit 2026-03-07T10:10:42.081721+0000 mon.a (mon.0) 300 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:43.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:42 vm10 bash[19283]: audit 2026-03-07T10:10:42.081721+0000 mon.a (mon.0) 300 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:43.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:42 vm02 bash[17011]: cluster 2026-03-07T10:10:40.802008+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T10:10:43.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:42 vm02 bash[17011]: cluster 2026-03-07T10:10:40.802008+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T10:10:43.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:42 vm02 bash[17011]: cluster 2026-03-07T10:10:40.802064+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T10:10:43.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:42 vm02 bash[17011]: cluster 2026-03-07T10:10:40.802064+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T10:10:43.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:42 vm02 bash[17011]: audit 2026-03-07T10:10:41.892222+0000 mon.a (mon.0) 292 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T10:10:43.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:42 vm02 bash[17011]: audit 2026-03-07T10:10:41.892222+0000 mon.a (mon.0) 292 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T10:10:43.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:42 vm02 bash[17011]: cluster 2026-03-07T10:10:41.905629+0000 mon.a (mon.0) 293 : cluster [INF] osd.1 [v2:192.168.123.110:6800/1483542873,v1:192.168.123.110:6801/1483542873] boot 2026-03-07T10:10:43.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:42 vm02 bash[17011]: cluster 2026-03-07T10:10:41.905629+0000 mon.a (mon.0) 293 : cluster [INF] osd.1 [v2:192.168.123.110:6800/1483542873,v1:192.168.123.110:6801/1483542873] boot 2026-03-07T10:10:43.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:42 vm02 bash[17011]: cluster 2026-03-07T10:10:41.905656+0000 mon.a (mon.0) 294 : cluster [DBG] osdmap e13: 2 total, 2 up, 2 in 2026-03-07T10:10:43.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:42 vm02 bash[17011]: cluster 2026-03-07T10:10:41.905656+0000 mon.a (mon.0) 294 : cluster [DBG] osdmap e13: 2 total, 2 up, 2 in 2026-03-07T10:10:43.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:42 vm02 bash[17011]: audit 2026-03-07T10:10:41.905701+0000 mon.a (mon.0) 295 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T10:10:43.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:42 vm02 bash[17011]: audit 2026-03-07T10:10:41.905701+0000 mon.a (mon.0) 295 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T10:10:43.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:42 vm02 bash[17011]: audit 2026-03-07T10:10:42.074598+0000 mon.a (mon.0) 296 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:43.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:42 vm02 bash[17011]: audit 2026-03-07T10:10:42.074598+0000 mon.a (mon.0) 296 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:43.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:42 vm02 bash[17011]: audit 2026-03-07T10:10:42.077780+0000 mon.a (mon.0) 297 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:43.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:42 vm02 bash[17011]: audit 2026-03-07T10:10:42.077780+0000 mon.a (mon.0) 297 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:43.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:42 vm02 bash[17011]: audit 2026-03-07T10:10:42.078310+0000 mon.a (mon.0) 298 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:10:43.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:42 vm02 bash[17011]: audit 2026-03-07T10:10:42.078310+0000 mon.a (mon.0) 298 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:10:43.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:42 vm02 bash[17011]: audit 2026-03-07T10:10:42.078692+0000 mon.a (mon.0) 299 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:10:43.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:42 vm02 bash[17011]: audit 2026-03-07T10:10:42.078692+0000 mon.a (mon.0) 299 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:10:43.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:42 vm02 bash[17011]: audit 2026-03-07T10:10:42.081721+0000 mon.a (mon.0) 300 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:43.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:42 vm02 bash[17011]: audit 2026-03-07T10:10:42.081721+0000 mon.a (mon.0) 300 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:43.496 INFO:teuthology.orchestra.run.vm10.stdout:Created osd(s) 1 on host 'vm10' 2026-03-07T10:10:43.608 DEBUG:teuthology.orchestra.run.vm10:osd.1> sudo journalctl -f -n 0 -u ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@osd.1.service 2026-03-07T10:10:43.609 INFO:tasks.cephadm:Waiting for 2 OSDs to come up... 2026-03-07T10:10:43.609 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph osd stat -f json 2026-03-07T10:10:44.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:43 vm10 bash[19283]: cluster 2026-03-07T10:10:42.280089+0000 mgr.a (mgr.14156) 99 : cluster [DBG] pgmap v66: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:44.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:43 vm10 bash[19283]: cluster 2026-03-07T10:10:42.280089+0000 mgr.a (mgr.14156) 99 : cluster [DBG] pgmap v66: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:44.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:43 vm10 bash[19283]: cluster 2026-03-07T10:10:43.087765+0000 mon.a (mon.0) 301 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-07T10:10:44.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:43 vm10 bash[19283]: cluster 2026-03-07T10:10:43.087765+0000 mon.a (mon.0) 301 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-07T10:10:44.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:43 vm10 bash[19283]: audit 2026-03-07T10:10:43.483123+0000 mon.a (mon.0) 302 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:10:44.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:43 vm10 bash[19283]: audit 2026-03-07T10:10:43.483123+0000 mon.a (mon.0) 302 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:10:44.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:43 vm10 bash[19283]: audit 2026-03-07T10:10:43.488147+0000 mon.a (mon.0) 303 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:44.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:43 vm10 bash[19283]: audit 2026-03-07T10:10:43.488147+0000 mon.a (mon.0) 303 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:44.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:43 vm10 bash[19283]: audit 2026-03-07T10:10:43.492235+0000 mon.a (mon.0) 304 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:44.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:43 vm10 bash[19283]: audit 2026-03-07T10:10:43.492235+0000 mon.a (mon.0) 304 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:44.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:43 vm02 bash[17011]: cluster 2026-03-07T10:10:42.280089+0000 mgr.a (mgr.14156) 99 : cluster [DBG] pgmap v66: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:44.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:43 vm02 bash[17011]: cluster 2026-03-07T10:10:42.280089+0000 mgr.a (mgr.14156) 99 : cluster [DBG] pgmap v66: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-07T10:10:44.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:43 vm02 bash[17011]: cluster 2026-03-07T10:10:43.087765+0000 mon.a (mon.0) 301 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-07T10:10:44.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:43 vm02 bash[17011]: cluster 2026-03-07T10:10:43.087765+0000 mon.a (mon.0) 301 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-07T10:10:44.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:43 vm02 bash[17011]: audit 2026-03-07T10:10:43.483123+0000 mon.a (mon.0) 302 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:10:44.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:43 vm02 bash[17011]: audit 2026-03-07T10:10:43.483123+0000 mon.a (mon.0) 302 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:10:44.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:43 vm02 bash[17011]: audit 2026-03-07T10:10:43.488147+0000 mon.a (mon.0) 303 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:44.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:43 vm02 bash[17011]: audit 2026-03-07T10:10:43.488147+0000 mon.a (mon.0) 303 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:44.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:43 vm02 bash[17011]: audit 2026-03-07T10:10:43.492235+0000 mon.a (mon.0) 304 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:44.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:43 vm02 bash[17011]: audit 2026-03-07T10:10:43.492235+0000 mon.a (mon.0) 304 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:46.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:45 vm10 bash[19283]: cluster 2026-03-07T10:10:44.280668+0000 mgr.a (mgr.14156) 100 : cluster [DBG] pgmap v68: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:46.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:45 vm10 bash[19283]: cluster 2026-03-07T10:10:44.280668+0000 mgr.a (mgr.14156) 100 : cluster [DBG] pgmap v68: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:46.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:45 vm02 bash[17011]: cluster 2026-03-07T10:10:44.280668+0000 mgr.a (mgr.14156) 100 : cluster [DBG] pgmap v68: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:46.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:45 vm02 bash[17011]: cluster 2026-03-07T10:10:44.280668+0000 mgr.a (mgr.14156) 100 : cluster [DBG] pgmap v68: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:48.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:47 vm10 bash[19283]: cluster 2026-03-07T10:10:46.281061+0000 mgr.a (mgr.14156) 101 : cluster [DBG] pgmap v69: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:48.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:47 vm10 bash[19283]: cluster 2026-03-07T10:10:46.281061+0000 mgr.a (mgr.14156) 101 : cluster [DBG] pgmap v69: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:48.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:47 vm02 bash[17011]: cluster 2026-03-07T10:10:46.281061+0000 mgr.a (mgr.14156) 101 : cluster [DBG] pgmap v69: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:48.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:47 vm02 bash[17011]: cluster 2026-03-07T10:10:46.281061+0000 mgr.a (mgr.14156) 101 : cluster [DBG] pgmap v69: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:48.374 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config 2026-03-07T10:10:48.712 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-07T10:10:48.775 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":14,"num_osds":2,"num_up_osds":2,"osd_up_since":1772878241,"num_in_osds":2,"osd_in_since":1772878224,"num_remapped_pgs":0} 2026-03-07T10:10:48.776 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph osd dump --format=json 2026-03-07T10:10:49.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:48 vm10 bash[19283]: audit 2026-03-07T10:10:48.711688+0000 mon.a (mon.0) 305 : audit [DBG] from='client.? 192.168.123.102:0/1925092013' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T10:10:49.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:48 vm10 bash[19283]: audit 2026-03-07T10:10:48.711688+0000 mon.a (mon.0) 305 : audit [DBG] from='client.? 192.168.123.102:0/1925092013' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T10:10:49.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:48 vm02 bash[17011]: audit 2026-03-07T10:10:48.711688+0000 mon.a (mon.0) 305 : audit [DBG] from='client.? 192.168.123.102:0/1925092013' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T10:10:49.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:48 vm02 bash[17011]: audit 2026-03-07T10:10:48.711688+0000 mon.a (mon.0) 305 : audit [DBG] from='client.? 192.168.123.102:0/1925092013' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T10:10:50.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:49 vm10 bash[19283]: cluster 2026-03-07T10:10:48.281327+0000 mgr.a (mgr.14156) 102 : cluster [DBG] pgmap v70: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:50.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:49 vm10 bash[19283]: cluster 2026-03-07T10:10:48.281327+0000 mgr.a (mgr.14156) 102 : cluster [DBG] pgmap v70: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:50.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:49 vm10 bash[19283]: audit 2026-03-07T10:10:49.405503+0000 mon.a (mon.0) 306 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:50.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:49 vm10 bash[19283]: audit 2026-03-07T10:10:49.405503+0000 mon.a (mon.0) 306 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:50.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:49 vm10 bash[19283]: audit 2026-03-07T10:10:49.410407+0000 mon.a (mon.0) 307 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:50.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:49 vm10 bash[19283]: audit 2026-03-07T10:10:49.410407+0000 mon.a (mon.0) 307 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:50.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:49 vm10 bash[19283]: audit 2026-03-07T10:10:49.411608+0000 mon.a (mon.0) 308 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-07T10:10:50.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:49 vm10 bash[19283]: audit 2026-03-07T10:10:49.411608+0000 mon.a (mon.0) 308 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-07T10:10:50.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:49 vm10 bash[19283]: audit 2026-03-07T10:10:49.412708+0000 mon.a (mon.0) 309 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:10:50.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:49 vm10 bash[19283]: audit 2026-03-07T10:10:49.412708+0000 mon.a (mon.0) 309 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:10:50.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:49 vm10 bash[19283]: audit 2026-03-07T10:10:49.413101+0000 mon.a (mon.0) 310 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:10:50.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:49 vm10 bash[19283]: audit 2026-03-07T10:10:49.413101+0000 mon.a (mon.0) 310 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:10:50.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:49 vm10 bash[19283]: audit 2026-03-07T10:10:49.416879+0000 mon.a (mon.0) 311 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:50.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:49 vm10 bash[19283]: audit 2026-03-07T10:10:49.416879+0000 mon.a (mon.0) 311 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:50.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:49 vm02 bash[17011]: cluster 2026-03-07T10:10:48.281327+0000 mgr.a (mgr.14156) 102 : cluster [DBG] pgmap v70: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:50.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:49 vm02 bash[17011]: cluster 2026-03-07T10:10:48.281327+0000 mgr.a (mgr.14156) 102 : cluster [DBG] pgmap v70: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:50.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:49 vm02 bash[17011]: audit 2026-03-07T10:10:49.405503+0000 mon.a (mon.0) 306 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:50.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:49 vm02 bash[17011]: audit 2026-03-07T10:10:49.405503+0000 mon.a (mon.0) 306 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:50.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:49 vm02 bash[17011]: audit 2026-03-07T10:10:49.410407+0000 mon.a (mon.0) 307 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:50.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:49 vm02 bash[17011]: audit 2026-03-07T10:10:49.410407+0000 mon.a (mon.0) 307 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:50.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:49 vm02 bash[17011]: audit 2026-03-07T10:10:49.411608+0000 mon.a (mon.0) 308 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-07T10:10:50.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:49 vm02 bash[17011]: audit 2026-03-07T10:10:49.411608+0000 mon.a (mon.0) 308 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-07T10:10:50.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:49 vm02 bash[17011]: audit 2026-03-07T10:10:49.412708+0000 mon.a (mon.0) 309 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:10:50.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:49 vm02 bash[17011]: audit 2026-03-07T10:10:49.412708+0000 mon.a (mon.0) 309 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:10:50.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:49 vm02 bash[17011]: audit 2026-03-07T10:10:49.413101+0000 mon.a (mon.0) 310 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:10:50.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:49 vm02 bash[17011]: audit 2026-03-07T10:10:49.413101+0000 mon.a (mon.0) 310 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:10:50.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:49 vm02 bash[17011]: audit 2026-03-07T10:10:49.416879+0000 mon.a (mon.0) 311 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:50.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:49 vm02 bash[17011]: audit 2026-03-07T10:10:49.416879+0000 mon.a (mon.0) 311 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:10:51.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:50 vm10 bash[19283]: cephadm 2026-03-07T10:10:49.399486+0000 mgr.a (mgr.14156) 103 : cephadm [INF] Detected new or changed devices on vm10 2026-03-07T10:10:51.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:50 vm10 bash[19283]: cephadm 2026-03-07T10:10:49.399486+0000 mgr.a (mgr.14156) 103 : cephadm [INF] Detected new or changed devices on vm10 2026-03-07T10:10:51.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:50 vm10 bash[19283]: cephadm 2026-03-07T10:10:49.412006+0000 mgr.a (mgr.14156) 104 : cephadm [INF] Adjusting osd_memory_target on vm10 to 455.7M 2026-03-07T10:10:51.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:50 vm10 bash[19283]: cephadm 2026-03-07T10:10:49.412006+0000 mgr.a (mgr.14156) 104 : cephadm [INF] Adjusting osd_memory_target on vm10 to 455.7M 2026-03-07T10:10:51.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:50 vm10 bash[19283]: cephadm 2026-03-07T10:10:49.412417+0000 mgr.a (mgr.14156) 105 : cephadm [WRN] Unable to set osd_memory_target on vm10 to 477915955: error parsing value: Value '477915955' is below minimum 939524096 2026-03-07T10:10:51.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:50 vm10 bash[19283]: cephadm 2026-03-07T10:10:49.412417+0000 mgr.a (mgr.14156) 105 : cephadm [WRN] Unable to set osd_memory_target on vm10 to 477915955: error parsing value: Value '477915955' is below minimum 939524096 2026-03-07T10:10:51.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:50 vm02 bash[17011]: cephadm 2026-03-07T10:10:49.399486+0000 mgr.a (mgr.14156) 103 : cephadm [INF] Detected new or changed devices on vm10 2026-03-07T10:10:51.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:50 vm02 bash[17011]: cephadm 2026-03-07T10:10:49.399486+0000 mgr.a (mgr.14156) 103 : cephadm [INF] Detected new or changed devices on vm10 2026-03-07T10:10:51.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:50 vm02 bash[17011]: cephadm 2026-03-07T10:10:49.412006+0000 mgr.a (mgr.14156) 104 : cephadm [INF] Adjusting osd_memory_target on vm10 to 455.7M 2026-03-07T10:10:51.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:50 vm02 bash[17011]: cephadm 2026-03-07T10:10:49.412006+0000 mgr.a (mgr.14156) 104 : cephadm [INF] Adjusting osd_memory_target on vm10 to 455.7M 2026-03-07T10:10:51.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:50 vm02 bash[17011]: cephadm 2026-03-07T10:10:49.412417+0000 mgr.a (mgr.14156) 105 : cephadm [WRN] Unable to set osd_memory_target on vm10 to 477915955: error parsing value: Value '477915955' is below minimum 939524096 2026-03-07T10:10:51.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:50 vm02 bash[17011]: cephadm 2026-03-07T10:10:49.412417+0000 mgr.a (mgr.14156) 105 : cephadm [WRN] Unable to set osd_memory_target on vm10 to 477915955: error parsing value: Value '477915955' is below minimum 939524096 2026-03-07T10:10:52.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:51 vm10 bash[19283]: cluster 2026-03-07T10:10:50.281566+0000 mgr.a (mgr.14156) 106 : cluster [DBG] pgmap v71: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:52.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:51 vm10 bash[19283]: cluster 2026-03-07T10:10:50.281566+0000 mgr.a (mgr.14156) 106 : cluster [DBG] pgmap v71: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:52.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:51 vm02 bash[17011]: cluster 2026-03-07T10:10:50.281566+0000 mgr.a (mgr.14156) 106 : cluster [DBG] pgmap v71: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:52.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:51 vm02 bash[17011]: cluster 2026-03-07T10:10:50.281566+0000 mgr.a (mgr.14156) 106 : cluster [DBG] pgmap v71: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:53.547 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config 2026-03-07T10:10:53.871 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-07T10:10:53.871 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":14,"fsid":"6c715b7a-1a0d-11f1-b180-89615ccd948e","created":"2026-03-07T10:07:53.158394+0000","modified":"2026-03-07T10:10:43.081686+0000","last_up_change":"2026-03-07T10:10:41.899964+0000","last_in_change":"2026-03-07T10:10:24.578023+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":6,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":0,"max_osd":2,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[],"osds":[{"osd":0,"uuid":"eb891bbe-0ffc-4a69-afbc-01132f549280","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6802","nonce":3942075313},{"type":"v1","addr":"192.168.123.102:6803","nonce":3942075313}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6804","nonce":3942075313},{"type":"v1","addr":"192.168.123.102:6805","nonce":3942075313}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6808","nonce":3942075313},{"type":"v1","addr":"192.168.123.102:6809","nonce":3942075313}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6806","nonce":3942075313},{"type":"v1","addr":"192.168.123.102:6807","nonce":3942075313}]},"public_addr":"192.168.123.102:6803/3942075313","cluster_addr":"192.168.123.102:6805/3942075313","heartbeat_back_addr":"192.168.123.102:6809/3942075313","heartbeat_front_addr":"192.168.123.102:6807/3942075313","state":["exists","up"]},{"osd":1,"uuid":"4d550476-452d-4382-8bdf-ae95a10d4527","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6800","nonce":1483542873},{"type":"v1","addr":"192.168.123.110:6801","nonce":1483542873}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6802","nonce":1483542873},{"type":"v1","addr":"192.168.123.110:6803","nonce":1483542873}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6806","nonce":1483542873},{"type":"v1","addr":"192.168.123.110:6807","nonce":1483542873}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6804","nonce":1483542873},{"type":"v1","addr":"192.168.123.110:6805","nonce":1483542873}]},"public_addr":"192.168.123.110:6801/1483542873","cluster_addr":"192.168.123.110:6803/1483542873","heartbeat_back_addr":"192.168.123.110:6807/1483542873","heartbeat_front_addr":"192.168.123.110:6805/1483542873","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T10:10:04.837869+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T10:10:40.802066+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.102:6801/2193623680":"2026-03-08T10:08:36.254066+0000","192.168.123.102:0/926547866":"2026-03-08T10:08:36.254066+0000","192.168.123.102:6800/2193623680":"2026-03-08T10:08:36.254066+0000","192.168.123.102:0/3799062051":"2026-03-08T10:08:36.254066+0000","192.168.123.102:0/2232612251":"2026-03-08T10:08:18.965575+0000","192.168.123.102:0/4096123239":"2026-03-08T10:08:18.965575+0000","192.168.123.102:0/3202277196":"2026-03-08T10:08:36.254066+0000","192.168.123.102:0/1717251641":"2026-03-08T10:08:18.965575+0000","192.168.123.102:6801/114762242":"2026-03-08T10:08:18.965575+0000","192.168.123.102:6800/114762242":"2026-03-08T10:08:18.965575+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-07T10:10:53.951 INFO:tasks.cephadm.ceph_manager.ceph:[] 2026-03-07T10:10:53.951 INFO:tasks.cephadm:Setting up client nodes... 2026-03-07T10:10:53.951 INFO:tasks.ceph:Waiting until ceph daemons up and pgs clean... 2026-03-07T10:10:53.951 INFO:tasks.cephadm.ceph_manager.ceph:waiting for mgr available 2026-03-07T10:10:53.951 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph mgr dump --format=json 2026-03-07T10:10:54.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:53 vm10 bash[19283]: cluster 2026-03-07T10:10:52.281774+0000 mgr.a (mgr.14156) 107 : cluster [DBG] pgmap v72: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:54.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:53 vm10 bash[19283]: cluster 2026-03-07T10:10:52.281774+0000 mgr.a (mgr.14156) 107 : cluster [DBG] pgmap v72: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:54.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:53 vm10 bash[19283]: audit 2026-03-07T10:10:53.870463+0000 mon.a (mon.0) 312 : audit [DBG] from='client.? 192.168.123.102:0/4267306391' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-07T10:10:54.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:53 vm10 bash[19283]: audit 2026-03-07T10:10:53.870463+0000 mon.a (mon.0) 312 : audit [DBG] from='client.? 192.168.123.102:0/4267306391' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-07T10:10:54.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:53 vm02 bash[17011]: cluster 2026-03-07T10:10:52.281774+0000 mgr.a (mgr.14156) 107 : cluster [DBG] pgmap v72: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:54.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:53 vm02 bash[17011]: cluster 2026-03-07T10:10:52.281774+0000 mgr.a (mgr.14156) 107 : cluster [DBG] pgmap v72: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:54.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:53 vm02 bash[17011]: audit 2026-03-07T10:10:53.870463+0000 mon.a (mon.0) 312 : audit [DBG] from='client.? 192.168.123.102:0/4267306391' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-07T10:10:54.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:53 vm02 bash[17011]: audit 2026-03-07T10:10:53.870463+0000 mon.a (mon.0) 312 : audit [DBG] from='client.? 192.168.123.102:0/4267306391' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-07T10:10:56.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:55 vm10 bash[19283]: cluster 2026-03-07T10:10:54.282112+0000 mgr.a (mgr.14156) 108 : cluster [DBG] pgmap v73: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:56.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:55 vm10 bash[19283]: cluster 2026-03-07T10:10:54.282112+0000 mgr.a (mgr.14156) 108 : cluster [DBG] pgmap v73: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:56.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:55 vm02 bash[17011]: cluster 2026-03-07T10:10:54.282112+0000 mgr.a (mgr.14156) 108 : cluster [DBG] pgmap v73: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:56.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:55 vm02 bash[17011]: cluster 2026-03-07T10:10:54.282112+0000 mgr.a (mgr.14156) 108 : cluster [DBG] pgmap v73: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:58.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:57 vm02 bash[17011]: cluster 2026-03-07T10:10:56.282358+0000 mgr.a (mgr.14156) 109 : cluster [DBG] pgmap v74: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:58.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:57 vm02 bash[17011]: cluster 2026-03-07T10:10:56.282358+0000 mgr.a (mgr.14156) 109 : cluster [DBG] pgmap v74: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:58.445 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:57 vm10 bash[19283]: cluster 2026-03-07T10:10:56.282358+0000 mgr.a (mgr.14156) 109 : cluster [DBG] pgmap v74: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:58.446 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:57 vm10 bash[19283]: cluster 2026-03-07T10:10:56.282358+0000 mgr.a (mgr.14156) 109 : cluster [DBG] pgmap v74: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:58.698 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config 2026-03-07T10:10:59.067 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-07T10:10:59.132 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":14,"flags":0,"active_gid":14156,"active_name":"a","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6800","nonce":3513675105},{"type":"v1","addr":"192.168.123.102:6801","nonce":3513675105}]},"active_addr":"192.168.123.102:6801/3513675105","active_change":"2026-03-07T10:08:36.254374+0000","active_mgr_features":4540701547738038271,"available":true,"standbys":[{"gid":24107,"name":"b","mgr_features":4540701547738038271,"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}]}],"modules":["cephadm","dashboard","iostat","nfs","restful"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{"dashboard":"https://192.168.123.102:8443/"},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"reef":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"squid":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"]},"force_disabled_modules":{},"last_failure_osd_epoch":3,"active_clients":[{"name":"libcephsqlite","addrvec":[{"type":"v2","addr":"192.168.123.102:0","nonce":825830265}]},{"name":"rbd_support","addrvec":[{"type":"v2","addr":"192.168.123.102:0","nonce":926838608}]},{"name":"volumes","addrvec":[{"type":"v2","addr":"192.168.123.102:0","nonce":325096336}]}]} 2026-03-07T10:10:59.134 INFO:tasks.cephadm.ceph_manager.ceph:mgr available! 2026-03-07T10:10:59.134 INFO:tasks.cephadm.ceph_manager.ceph:waiting for all up 2026-03-07T10:10:59.134 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph osd dump --format=json 2026-03-07T10:10:59.961 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:59 vm02 bash[17011]: cluster 2026-03-07T10:10:58.282568+0000 mgr.a (mgr.14156) 110 : cluster [DBG] pgmap v75: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:59.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:59 vm02 bash[17011]: cluster 2026-03-07T10:10:58.282568+0000 mgr.a (mgr.14156) 110 : cluster [DBG] pgmap v75: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:10:59.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:59 vm02 bash[17011]: audit 2026-03-07T10:10:59.064308+0000 mon.a (mon.0) 313 : audit [DBG] from='client.? 192.168.123.102:0/2809566053' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-07T10:10:59.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:10:59 vm02 bash[17011]: audit 2026-03-07T10:10:59.064308+0000 mon.a (mon.0) 313 : audit [DBG] from='client.? 192.168.123.102:0/2809566053' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-07T10:11:00.446 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:59 vm10 bash[19283]: cluster 2026-03-07T10:10:58.282568+0000 mgr.a (mgr.14156) 110 : cluster [DBG] pgmap v75: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:00.446 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:59 vm10 bash[19283]: cluster 2026-03-07T10:10:58.282568+0000 mgr.a (mgr.14156) 110 : cluster [DBG] pgmap v75: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:00.446 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:59 vm10 bash[19283]: audit 2026-03-07T10:10:59.064308+0000 mon.a (mon.0) 313 : audit [DBG] from='client.? 192.168.123.102:0/2809566053' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-07T10:11:00.446 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:10:59 vm10 bash[19283]: audit 2026-03-07T10:10:59.064308+0000 mon.a (mon.0) 313 : audit [DBG] from='client.? 192.168.123.102:0/2809566053' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-07T10:11:02.445 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:01 vm10 bash[19283]: cluster 2026-03-07T10:11:00.282791+0000 mgr.a (mgr.14156) 111 : cluster [DBG] pgmap v76: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:02.445 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:01 vm10 bash[19283]: cluster 2026-03-07T10:11:00.282791+0000 mgr.a (mgr.14156) 111 : cluster [DBG] pgmap v76: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:02.461 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:01 vm02 bash[17011]: cluster 2026-03-07T10:11:00.282791+0000 mgr.a (mgr.14156) 111 : cluster [DBG] pgmap v76: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:02.461 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:01 vm02 bash[17011]: cluster 2026-03-07T10:11:00.282791+0000 mgr.a (mgr.14156) 111 : cluster [DBG] pgmap v76: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:03.892 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config 2026-03-07T10:11:04.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:03 vm02 bash[17011]: cluster 2026-03-07T10:11:02.283017+0000 mgr.a (mgr.14156) 112 : cluster [DBG] pgmap v77: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:04.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:03 vm02 bash[17011]: cluster 2026-03-07T10:11:02.283017+0000 mgr.a (mgr.14156) 112 : cluster [DBG] pgmap v77: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:04.236 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-07T10:11:04.236 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":14,"fsid":"6c715b7a-1a0d-11f1-b180-89615ccd948e","created":"2026-03-07T10:07:53.158394+0000","modified":"2026-03-07T10:10:43.081686+0000","last_up_change":"2026-03-07T10:10:41.899964+0000","last_in_change":"2026-03-07T10:10:24.578023+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":6,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":0,"max_osd":2,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[],"osds":[{"osd":0,"uuid":"eb891bbe-0ffc-4a69-afbc-01132f549280","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6802","nonce":3942075313},{"type":"v1","addr":"192.168.123.102:6803","nonce":3942075313}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6804","nonce":3942075313},{"type":"v1","addr":"192.168.123.102:6805","nonce":3942075313}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6808","nonce":3942075313},{"type":"v1","addr":"192.168.123.102:6809","nonce":3942075313}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6806","nonce":3942075313},{"type":"v1","addr":"192.168.123.102:6807","nonce":3942075313}]},"public_addr":"192.168.123.102:6803/3942075313","cluster_addr":"192.168.123.102:6805/3942075313","heartbeat_back_addr":"192.168.123.102:6809/3942075313","heartbeat_front_addr":"192.168.123.102:6807/3942075313","state":["exists","up"]},{"osd":1,"uuid":"4d550476-452d-4382-8bdf-ae95a10d4527","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6800","nonce":1483542873},{"type":"v1","addr":"192.168.123.110:6801","nonce":1483542873}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6802","nonce":1483542873},{"type":"v1","addr":"192.168.123.110:6803","nonce":1483542873}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6806","nonce":1483542873},{"type":"v1","addr":"192.168.123.110:6807","nonce":1483542873}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6804","nonce":1483542873},{"type":"v1","addr":"192.168.123.110:6805","nonce":1483542873}]},"public_addr":"192.168.123.110:6801/1483542873","cluster_addr":"192.168.123.110:6803/1483542873","heartbeat_back_addr":"192.168.123.110:6807/1483542873","heartbeat_front_addr":"192.168.123.110:6805/1483542873","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T10:10:04.837869+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T10:10:40.802066+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.102:6801/2193623680":"2026-03-08T10:08:36.254066+0000","192.168.123.102:0/926547866":"2026-03-08T10:08:36.254066+0000","192.168.123.102:6800/2193623680":"2026-03-08T10:08:36.254066+0000","192.168.123.102:0/3799062051":"2026-03-08T10:08:36.254066+0000","192.168.123.102:0/2232612251":"2026-03-08T10:08:18.965575+0000","192.168.123.102:0/4096123239":"2026-03-08T10:08:18.965575+0000","192.168.123.102:0/3202277196":"2026-03-08T10:08:36.254066+0000","192.168.123.102:0/1717251641":"2026-03-08T10:08:18.965575+0000","192.168.123.102:6801/114762242":"2026-03-08T10:08:18.965575+0000","192.168.123.102:6800/114762242":"2026-03-08T10:08:18.965575+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-07T10:11:04.293 INFO:tasks.cephadm.ceph_manager.ceph:all up! 2026-03-07T10:11:04.293 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph osd dump --format=json 2026-03-07T10:11:04.445 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:03 vm10 bash[19283]: cluster 2026-03-07T10:11:02.283017+0000 mgr.a (mgr.14156) 112 : cluster [DBG] pgmap v77: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:04.445 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:03 vm10 bash[19283]: cluster 2026-03-07T10:11:02.283017+0000 mgr.a (mgr.14156) 112 : cluster [DBG] pgmap v77: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:05.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:04 vm02 bash[17011]: audit 2026-03-07T10:11:04.235014+0000 mon.a (mon.0) 314 : audit [DBG] from='client.? 192.168.123.102:0/3834554940' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-07T10:11:05.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:04 vm02 bash[17011]: audit 2026-03-07T10:11:04.235014+0000 mon.a (mon.0) 314 : audit [DBG] from='client.? 192.168.123.102:0/3834554940' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-07T10:11:05.445 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:04 vm10 bash[19283]: audit 2026-03-07T10:11:04.235014+0000 mon.a (mon.0) 314 : audit [DBG] from='client.? 192.168.123.102:0/3834554940' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-07T10:11:05.445 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:04 vm10 bash[19283]: audit 2026-03-07T10:11:04.235014+0000 mon.a (mon.0) 314 : audit [DBG] from='client.? 192.168.123.102:0/3834554940' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-07T10:11:06.445 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:05 vm10 bash[19283]: cluster 2026-03-07T10:11:04.283262+0000 mgr.a (mgr.14156) 113 : cluster [DBG] pgmap v78: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:06.445 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:05 vm10 bash[19283]: cluster 2026-03-07T10:11:04.283262+0000 mgr.a (mgr.14156) 113 : cluster [DBG] pgmap v78: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:06.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:05 vm02 bash[17011]: cluster 2026-03-07T10:11:04.283262+0000 mgr.a (mgr.14156) 113 : cluster [DBG] pgmap v78: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:06.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:05 vm02 bash[17011]: cluster 2026-03-07T10:11:04.283262+0000 mgr.a (mgr.14156) 113 : cluster [DBG] pgmap v78: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:08.445 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:07 vm10 bash[19283]: cluster 2026-03-07T10:11:06.283512+0000 mgr.a (mgr.14156) 114 : cluster [DBG] pgmap v79: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:08.446 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:07 vm10 bash[19283]: cluster 2026-03-07T10:11:06.283512+0000 mgr.a (mgr.14156) 114 : cluster [DBG] pgmap v79: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:08.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:07 vm02 bash[17011]: cluster 2026-03-07T10:11:06.283512+0000 mgr.a (mgr.14156) 114 : cluster [DBG] pgmap v79: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:08.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:07 vm02 bash[17011]: cluster 2026-03-07T10:11:06.283512+0000 mgr.a (mgr.14156) 114 : cluster [DBG] pgmap v79: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:09.045 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config 2026-03-07T10:11:09.401 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-07T10:11:09.401 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":14,"fsid":"6c715b7a-1a0d-11f1-b180-89615ccd948e","created":"2026-03-07T10:07:53.158394+0000","modified":"2026-03-07T10:10:43.081686+0000","last_up_change":"2026-03-07T10:10:41.899964+0000","last_in_change":"2026-03-07T10:10:24.578023+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":6,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":0,"max_osd":2,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[],"osds":[{"osd":0,"uuid":"eb891bbe-0ffc-4a69-afbc-01132f549280","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6802","nonce":3942075313},{"type":"v1","addr":"192.168.123.102:6803","nonce":3942075313}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6804","nonce":3942075313},{"type":"v1","addr":"192.168.123.102:6805","nonce":3942075313}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6808","nonce":3942075313},{"type":"v1","addr":"192.168.123.102:6809","nonce":3942075313}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6806","nonce":3942075313},{"type":"v1","addr":"192.168.123.102:6807","nonce":3942075313}]},"public_addr":"192.168.123.102:6803/3942075313","cluster_addr":"192.168.123.102:6805/3942075313","heartbeat_back_addr":"192.168.123.102:6809/3942075313","heartbeat_front_addr":"192.168.123.102:6807/3942075313","state":["exists","up"]},{"osd":1,"uuid":"4d550476-452d-4382-8bdf-ae95a10d4527","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6800","nonce":1483542873},{"type":"v1","addr":"192.168.123.110:6801","nonce":1483542873}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6802","nonce":1483542873},{"type":"v1","addr":"192.168.123.110:6803","nonce":1483542873}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6806","nonce":1483542873},{"type":"v1","addr":"192.168.123.110:6807","nonce":1483542873}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6804","nonce":1483542873},{"type":"v1","addr":"192.168.123.110:6805","nonce":1483542873}]},"public_addr":"192.168.123.110:6801/1483542873","cluster_addr":"192.168.123.110:6803/1483542873","heartbeat_back_addr":"192.168.123.110:6807/1483542873","heartbeat_front_addr":"192.168.123.110:6805/1483542873","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T10:10:04.837869+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T10:10:40.802066+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.102:6801/2193623680":"2026-03-08T10:08:36.254066+0000","192.168.123.102:0/926547866":"2026-03-08T10:08:36.254066+0000","192.168.123.102:6800/2193623680":"2026-03-08T10:08:36.254066+0000","192.168.123.102:0/3799062051":"2026-03-08T10:08:36.254066+0000","192.168.123.102:0/2232612251":"2026-03-08T10:08:18.965575+0000","192.168.123.102:0/4096123239":"2026-03-08T10:08:18.965575+0000","192.168.123.102:0/3202277196":"2026-03-08T10:08:36.254066+0000","192.168.123.102:0/1717251641":"2026-03-08T10:08:18.965575+0000","192.168.123.102:6801/114762242":"2026-03-08T10:08:18.965575+0000","192.168.123.102:6800/114762242":"2026-03-08T10:08:18.965575+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-07T10:11:09.468 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph tell osd.0 flush_pg_stats 2026-03-07T10:11:09.468 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph tell osd.1 flush_pg_stats 2026-03-07T10:11:10.445 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:10 vm10 bash[19283]: cluster 2026-03-07T10:11:08.283863+0000 mgr.a (mgr.14156) 115 : cluster [DBG] pgmap v80: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:10.446 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:10 vm10 bash[19283]: cluster 2026-03-07T10:11:08.283863+0000 mgr.a (mgr.14156) 115 : cluster [DBG] pgmap v80: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:10.446 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:10 vm10 bash[19283]: audit 2026-03-07T10:11:09.400359+0000 mon.a (mon.0) 315 : audit [DBG] from='client.? 192.168.123.102:0/3134453136' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-07T10:11:10.446 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:10 vm10 bash[19283]: audit 2026-03-07T10:11:09.400359+0000 mon.a (mon.0) 315 : audit [DBG] from='client.? 192.168.123.102:0/3134453136' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-07T10:11:10.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:10 vm02 bash[17011]: cluster 2026-03-07T10:11:08.283863+0000 mgr.a (mgr.14156) 115 : cluster [DBG] pgmap v80: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:10.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:10 vm02 bash[17011]: cluster 2026-03-07T10:11:08.283863+0000 mgr.a (mgr.14156) 115 : cluster [DBG] pgmap v80: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:10.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:10 vm02 bash[17011]: audit 2026-03-07T10:11:09.400359+0000 mon.a (mon.0) 315 : audit [DBG] from='client.? 192.168.123.102:0/3134453136' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-07T10:11:10.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:10 vm02 bash[17011]: audit 2026-03-07T10:11:09.400359+0000 mon.a (mon.0) 315 : audit [DBG] from='client.? 192.168.123.102:0/3134453136' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-07T10:11:11.445 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:11 vm10 bash[19283]: cluster 2026-03-07T10:11:10.284188+0000 mgr.a (mgr.14156) 116 : cluster [DBG] pgmap v81: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:11.445 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:11 vm10 bash[19283]: cluster 2026-03-07T10:11:10.284188+0000 mgr.a (mgr.14156) 116 : cluster [DBG] pgmap v81: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:11.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:11 vm02 bash[17011]: cluster 2026-03-07T10:11:10.284188+0000 mgr.a (mgr.14156) 116 : cluster [DBG] pgmap v81: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:11.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:11 vm02 bash[17011]: cluster 2026-03-07T10:11:10.284188+0000 mgr.a (mgr.14156) 116 : cluster [DBG] pgmap v81: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:13.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:13 vm10 bash[19283]: cluster 2026-03-07T10:11:12.284420+0000 mgr.a (mgr.14156) 117 : cluster [DBG] pgmap v82: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:13.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:13 vm10 bash[19283]: cluster 2026-03-07T10:11:12.284420+0000 mgr.a (mgr.14156) 117 : cluster [DBG] pgmap v82: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:13.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:13 vm02 bash[17011]: cluster 2026-03-07T10:11:12.284420+0000 mgr.a (mgr.14156) 117 : cluster [DBG] pgmap v82: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:13.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:13 vm02 bash[17011]: cluster 2026-03-07T10:11:12.284420+0000 mgr.a (mgr.14156) 117 : cluster [DBG] pgmap v82: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:14.247 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config 2026-03-07T10:11:14.249 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config 2026-03-07T10:11:14.578 INFO:teuthology.orchestra.run.vm02.stdout:34359738383 2026-03-07T10:11:14.578 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph osd last-stat-seq osd.0 2026-03-07T10:11:14.633 INFO:teuthology.orchestra.run.vm02.stdout:55834574856 2026-03-07T10:11:14.633 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph osd last-stat-seq osd.1 2026-03-07T10:11:15.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:15 vm10 bash[19283]: cluster 2026-03-07T10:11:14.284617+0000 mgr.a (mgr.14156) 118 : cluster [DBG] pgmap v83: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:15.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:15 vm10 bash[19283]: cluster 2026-03-07T10:11:14.284617+0000 mgr.a (mgr.14156) 118 : cluster [DBG] pgmap v83: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:15.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:15 vm02 bash[17011]: cluster 2026-03-07T10:11:14.284617+0000 mgr.a (mgr.14156) 118 : cluster [DBG] pgmap v83: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:15.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:15 vm02 bash[17011]: cluster 2026-03-07T10:11:14.284617+0000 mgr.a (mgr.14156) 118 : cluster [DBG] pgmap v83: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:17.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:17 vm10 bash[19283]: cluster 2026-03-07T10:11:16.284806+0000 mgr.a (mgr.14156) 119 : cluster [DBG] pgmap v84: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:17.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:17 vm10 bash[19283]: cluster 2026-03-07T10:11:16.284806+0000 mgr.a (mgr.14156) 119 : cluster [DBG] pgmap v84: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:17.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:17 vm02 bash[17011]: cluster 2026-03-07T10:11:16.284806+0000 mgr.a (mgr.14156) 119 : cluster [DBG] pgmap v84: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:17.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:17 vm02 bash[17011]: cluster 2026-03-07T10:11:16.284806+0000 mgr.a (mgr.14156) 119 : cluster [DBG] pgmap v84: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:19.343 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config 2026-03-07T10:11:19.345 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config 2026-03-07T10:11:19.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:19 vm10 bash[19283]: cluster 2026-03-07T10:11:18.285005+0000 mgr.a (mgr.14156) 120 : cluster [DBG] pgmap v85: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:19.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:19 vm10 bash[19283]: cluster 2026-03-07T10:11:18.285005+0000 mgr.a (mgr.14156) 120 : cluster [DBG] pgmap v85: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:19.698 INFO:teuthology.orchestra.run.vm02.stdout:34359738384 2026-03-07T10:11:19.698 INFO:teuthology.orchestra.run.vm02.stdout:55834574857 2026-03-07T10:11:19.709 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:19 vm02 bash[17011]: cluster 2026-03-07T10:11:18.285005+0000 mgr.a (mgr.14156) 120 : cluster [DBG] pgmap v85: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:19.709 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:19 vm02 bash[17011]: cluster 2026-03-07T10:11:18.285005+0000 mgr.a (mgr.14156) 120 : cluster [DBG] pgmap v85: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:19.783 INFO:tasks.cephadm.ceph_manager.ceph:need seq 34359738383 got 34359738384 for osd.0 2026-03-07T10:11:19.783 DEBUG:teuthology.parallel:result is None 2026-03-07T10:11:19.805 INFO:tasks.cephadm.ceph_manager.ceph:need seq 55834574856 got 55834574857 for osd.1 2026-03-07T10:11:19.805 DEBUG:teuthology.parallel:result is None 2026-03-07T10:11:19.805 INFO:tasks.cephadm.ceph_manager.ceph:waiting for clean 2026-03-07T10:11:19.805 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph pg dump --format=json 2026-03-07T10:11:20.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:20 vm10 bash[19283]: audit 2026-03-07T10:11:19.696375+0000 mon.a (mon.0) 316 : audit [DBG] from='client.? 192.168.123.102:0/727732558' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-07T10:11:20.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:20 vm10 bash[19283]: audit 2026-03-07T10:11:19.696375+0000 mon.a (mon.0) 316 : audit [DBG] from='client.? 192.168.123.102:0/727732558' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-07T10:11:20.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:20 vm10 bash[19283]: audit 2026-03-07T10:11:19.697814+0000 mon.a (mon.0) 317 : audit [DBG] from='client.? 192.168.123.102:0/2816492529' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-07T10:11:20.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:20 vm10 bash[19283]: audit 2026-03-07T10:11:19.697814+0000 mon.a (mon.0) 317 : audit [DBG] from='client.? 192.168.123.102:0/2816492529' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-07T10:11:20.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:20 vm02 bash[17011]: audit 2026-03-07T10:11:19.696375+0000 mon.a (mon.0) 316 : audit [DBG] from='client.? 192.168.123.102:0/727732558' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-07T10:11:20.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:20 vm02 bash[17011]: audit 2026-03-07T10:11:19.696375+0000 mon.a (mon.0) 316 : audit [DBG] from='client.? 192.168.123.102:0/727732558' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-07T10:11:20.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:20 vm02 bash[17011]: audit 2026-03-07T10:11:19.697814+0000 mon.a (mon.0) 317 : audit [DBG] from='client.? 192.168.123.102:0/2816492529' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-07T10:11:20.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:20 vm02 bash[17011]: audit 2026-03-07T10:11:19.697814+0000 mon.a (mon.0) 317 : audit [DBG] from='client.? 192.168.123.102:0/2816492529' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-07T10:11:21.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:21 vm10 bash[19283]: cluster 2026-03-07T10:11:20.285273+0000 mgr.a (mgr.14156) 121 : cluster [DBG] pgmap v86: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:21.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:21 vm10 bash[19283]: cluster 2026-03-07T10:11:20.285273+0000 mgr.a (mgr.14156) 121 : cluster [DBG] pgmap v86: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:21.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:21 vm02 bash[17011]: cluster 2026-03-07T10:11:20.285273+0000 mgr.a (mgr.14156) 121 : cluster [DBG] pgmap v86: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:21.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:21 vm02 bash[17011]: cluster 2026-03-07T10:11:20.285273+0000 mgr.a (mgr.14156) 121 : cluster [DBG] pgmap v86: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:23.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:23 vm10 bash[19283]: cluster 2026-03-07T10:11:22.285484+0000 mgr.a (mgr.14156) 122 : cluster [DBG] pgmap v87: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:23.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:23 vm10 bash[19283]: cluster 2026-03-07T10:11:22.285484+0000 mgr.a (mgr.14156) 122 : cluster [DBG] pgmap v87: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:23.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:23 vm02 bash[17011]: cluster 2026-03-07T10:11:22.285484+0000 mgr.a (mgr.14156) 122 : cluster [DBG] pgmap v87: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:23.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:23 vm02 bash[17011]: cluster 2026-03-07T10:11:22.285484+0000 mgr.a (mgr.14156) 122 : cluster [DBG] pgmap v87: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:24.610 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config 2026-03-07T10:11:24.947 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-07T10:11:24.948 INFO:teuthology.orchestra.run.vm02.stderr:dumped all 2026-03-07T10:11:25.004 INFO:teuthology.orchestra.run.vm02.stdout:{"pg_ready":true,"pg_map":{"version":88,"stamp":"2026-03-07T10:11:24.285633+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":0,"num_osds":2,"num_per_pool_osds":2,"num_per_pool_omap_osds":0,"kb":41934848,"kb_used":53924,"kb_used_data":240,"kb_used_omap":3,"kb_used_meta":53628,"kb_avail":41880924,"statfs":{"total":42941284352,"available":42886066176,"internally_reserved":0,"allocated":245760,"data_stored":60148,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":3180,"internal_metadata":54915988},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"0.000000"},"pg_stats":[],"pool_stats":[],"osd_stats":[{"osd":1,"up_from":13,"seq":55834574858,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":26960,"kb_used_data":120,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940464,"statfs":{"total":21470642176,"available":21443035136,"internally_reserved":0,"allocated":122880,"data_stored":30074,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":8,"seq":34359738385,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":26964,"kb_used_data":120,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940460,"statfs":{"total":21470642176,"available":21443031040,"internally_reserved":0,"allocated":122880,"data_stored":30074,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[]}} 2026-03-07T10:11:25.004 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph pg dump --format=json 2026-03-07T10:11:25.385 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:25 vm02 bash[17011]: cluster 2026-03-07T10:11:24.285705+0000 mgr.a (mgr.14156) 123 : cluster [DBG] pgmap v88: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:25.385 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:25 vm02 bash[17011]: cluster 2026-03-07T10:11:24.285705+0000 mgr.a (mgr.14156) 123 : cluster [DBG] pgmap v88: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:25.385 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:25 vm02 bash[17011]: audit 2026-03-07T10:11:24.946885+0000 mgr.a (mgr.14156) 124 : audit [DBG] from='client.14268 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T10:11:25.385 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:25 vm02 bash[17011]: audit 2026-03-07T10:11:24.946885+0000 mgr.a (mgr.14156) 124 : audit [DBG] from='client.14268 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T10:11:25.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:25 vm10 bash[19283]: cluster 2026-03-07T10:11:24.285705+0000 mgr.a (mgr.14156) 123 : cluster [DBG] pgmap v88: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:25.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:25 vm10 bash[19283]: cluster 2026-03-07T10:11:24.285705+0000 mgr.a (mgr.14156) 123 : cluster [DBG] pgmap v88: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:25.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:25 vm10 bash[19283]: audit 2026-03-07T10:11:24.946885+0000 mgr.a (mgr.14156) 124 : audit [DBG] from='client.14268 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T10:11:25.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:25 vm10 bash[19283]: audit 2026-03-07T10:11:24.946885+0000 mgr.a (mgr.14156) 124 : audit [DBG] from='client.14268 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T10:11:27.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:27 vm10 bash[19283]: cluster 2026-03-07T10:11:26.285918+0000 mgr.a (mgr.14156) 125 : cluster [DBG] pgmap v89: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:27.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:27 vm10 bash[19283]: cluster 2026-03-07T10:11:26.285918+0000 mgr.a (mgr.14156) 125 : cluster [DBG] pgmap v89: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:27.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:27 vm02 bash[17011]: cluster 2026-03-07T10:11:26.285918+0000 mgr.a (mgr.14156) 125 : cluster [DBG] pgmap v89: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:27.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:27 vm02 bash[17011]: cluster 2026-03-07T10:11:26.285918+0000 mgr.a (mgr.14156) 125 : cluster [DBG] pgmap v89: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:29.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:29 vm10 bash[19283]: cluster 2026-03-07T10:11:28.286132+0000 mgr.a (mgr.14156) 126 : cluster [DBG] pgmap v90: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:29.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:29 vm10 bash[19283]: cluster 2026-03-07T10:11:28.286132+0000 mgr.a (mgr.14156) 126 : cluster [DBG] pgmap v90: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:29.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:29 vm02 bash[17011]: cluster 2026-03-07T10:11:28.286132+0000 mgr.a (mgr.14156) 126 : cluster [DBG] pgmap v90: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:29.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:29 vm02 bash[17011]: cluster 2026-03-07T10:11:28.286132+0000 mgr.a (mgr.14156) 126 : cluster [DBG] pgmap v90: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:29.760 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config 2026-03-07T10:11:30.138 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-07T10:11:30.138 INFO:teuthology.orchestra.run.vm02.stderr:dumped all 2026-03-07T10:11:30.198 INFO:teuthology.orchestra.run.vm02.stdout:{"pg_ready":true,"pg_map":{"version":90,"stamp":"2026-03-07T10:11:28.286053+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":0,"num_osds":2,"num_per_pool_osds":2,"num_per_pool_omap_osds":0,"kb":41934848,"kb_used":53924,"kb_used_data":240,"kb_used_omap":3,"kb_used_meta":53628,"kb_avail":41880924,"statfs":{"total":42941284352,"available":42886066176,"internally_reserved":0,"allocated":245760,"data_stored":60148,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":3180,"internal_metadata":54915988},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"0.000000"},"pg_stats":[],"pool_stats":[],"osd_stats":[{"osd":1,"up_from":13,"seq":55834574859,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":26960,"kb_used_data":120,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940464,"statfs":{"total":21470642176,"available":21443035136,"internally_reserved":0,"allocated":122880,"data_stored":30074,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":8,"seq":34359738386,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":26964,"kb_used_data":120,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940460,"statfs":{"total":21470642176,"available":21443031040,"internally_reserved":0,"allocated":122880,"data_stored":30074,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[]}} 2026-03-07T10:11:30.198 INFO:tasks.cephadm.ceph_manager.ceph:clean! 2026-03-07T10:11:30.198 INFO:tasks.ceph:Waiting until ceph cluster ceph is healthy... 2026-03-07T10:11:30.198 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy 2026-03-07T10:11:30.198 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph health --format=json 2026-03-07T10:11:31.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:31 vm10 bash[19283]: audit 2026-03-07T10:11:30.137338+0000 mgr.a (mgr.14156) 127 : audit [DBG] from='client.14272 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T10:11:31.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:31 vm10 bash[19283]: audit 2026-03-07T10:11:30.137338+0000 mgr.a (mgr.14156) 127 : audit [DBG] from='client.14272 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T10:11:31.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:31 vm10 bash[19283]: cluster 2026-03-07T10:11:30.286355+0000 mgr.a (mgr.14156) 128 : cluster [DBG] pgmap v91: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:31.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:31 vm10 bash[19283]: cluster 2026-03-07T10:11:30.286355+0000 mgr.a (mgr.14156) 128 : cluster [DBG] pgmap v91: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:31.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:31 vm02 bash[17011]: audit 2026-03-07T10:11:30.137338+0000 mgr.a (mgr.14156) 127 : audit [DBG] from='client.14272 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T10:11:31.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:31 vm02 bash[17011]: audit 2026-03-07T10:11:30.137338+0000 mgr.a (mgr.14156) 127 : audit [DBG] from='client.14272 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T10:11:31.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:31 vm02 bash[17011]: cluster 2026-03-07T10:11:30.286355+0000 mgr.a (mgr.14156) 128 : cluster [DBG] pgmap v91: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:31.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:31 vm02 bash[17011]: cluster 2026-03-07T10:11:30.286355+0000 mgr.a (mgr.14156) 128 : cluster [DBG] pgmap v91: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:33.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:33 vm10 bash[19283]: cluster 2026-03-07T10:11:32.286564+0000 mgr.a (mgr.14156) 129 : cluster [DBG] pgmap v92: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:33.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:33 vm10 bash[19283]: cluster 2026-03-07T10:11:32.286564+0000 mgr.a (mgr.14156) 129 : cluster [DBG] pgmap v92: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:33.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:33 vm02 bash[17011]: cluster 2026-03-07T10:11:32.286564+0000 mgr.a (mgr.14156) 129 : cluster [DBG] pgmap v92: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:33.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:33 vm02 bash[17011]: cluster 2026-03-07T10:11:32.286564+0000 mgr.a (mgr.14156) 129 : cluster [DBG] pgmap v92: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:34.969 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config 2026-03-07T10:11:35.330 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-07T10:11:35.330 INFO:teuthology.orchestra.run.vm02.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-03-07T10:11:35.396 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy done 2026-03-07T10:11:35.396 INFO:tasks.cephadm:Setup complete, yielding 2026-03-07T10:11:35.396 INFO:teuthology.run_tasks:Running task exec... 2026-03-07T10:11:35.399 INFO:teuthology.task.exec:Executing custom commands... 2026-03-07T10:11:35.399 INFO:teuthology.task.exec:Running commands on role host.a host ubuntu@vm02.local 2026-03-07T10:11:35.399 DEBUG:teuthology.orchestra.run.vm02:> sudo TESTDIR=/home/ubuntu/cephtest bash -c 'mkdir /etc/cephadm_testing' 2026-03-07T10:11:35.407 INFO:teuthology.task.exec:Running commands on role host.b host ubuntu@vm10.local 2026-03-07T10:11:35.407 DEBUG:teuthology.orchestra.run.vm10:> sudo TESTDIR=/home/ubuntu/cephtest bash -c 'mkdir /etc/cephadm_testing' 2026-03-07T10:11:35.415 INFO:teuthology.run_tasks:Running task cephadm.apply... 2026-03-07T10:11:35.418 INFO:tasks.cephadm:Applying spec(s): extra_container_args: - --cpus=2 extra_entrypoint_args: - --debug_ms 10 placement: host_pattern: '*' service_type: mon --- custom_configs: - content: "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n\ \ ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\n\ sleep infinity\n" mount_path: /root/write_thing_to_file.sh extra_container_args: - -v - /etc/cephadm_testing:/root/cephadm_testing extra_entrypoint_args: - /root/write_thing_to_file.sh - -c - testing_custom_containers - -o - /root/cephadm_testing/testing.txt placement: host_pattern: '*' service_id: foo service_type: container spec: entrypoint: bash image: quay.io/fedora/fedora:latest --- custom_configs: - content: 'set -e test -f /var/cache/bar/from.txt test -f /var/cache/bar/presized.dat echo ok > /var/cache/bar/primary.txt sleep infinity ' mount_path: /root/init_check.sh extra_entrypoint_args: - /root/init_check.sh placement: host_pattern: '*' service_id: bar service_type: container spec: dirs: - data entrypoint: bash image: quay.io/fedora/fedora:latest init_containers: - entrypoint: bash entrypoint_args: - argument: -c - argument: . /etc/os-release && echo from=$ID > /var/cache/bar/from.txt image: quay.io/centos/centos:latest volume_mounts: data: /var/cache/bar:z - entrypoint: bash entrypoint_args: - argument: -c - argument: test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat volume_mounts: data: /var/cache/bar:z volume_mounts: data: /var/cache/bar:z 2026-03-07T10:11:35.418 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph orch apply -i - 2026-03-07T10:11:35.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:35 vm10 bash[19283]: cluster 2026-03-07T10:11:34.286763+0000 mgr.a (mgr.14156) 130 : cluster [DBG] pgmap v93: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:35.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:35 vm10 bash[19283]: cluster 2026-03-07T10:11:34.286763+0000 mgr.a (mgr.14156) 130 : cluster [DBG] pgmap v93: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:35.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:35 vm10 bash[19283]: audit 2026-03-07T10:11:35.329211+0000 mon.a (mon.0) 318 : audit [DBG] from='client.? 192.168.123.102:0/336676369' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-07T10:11:35.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:35 vm10 bash[19283]: audit 2026-03-07T10:11:35.329211+0000 mon.a (mon.0) 318 : audit [DBG] from='client.? 192.168.123.102:0/336676369' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-07T10:11:35.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:35 vm02 bash[17011]: cluster 2026-03-07T10:11:34.286763+0000 mgr.a (mgr.14156) 130 : cluster [DBG] pgmap v93: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:35.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:35 vm02 bash[17011]: cluster 2026-03-07T10:11:34.286763+0000 mgr.a (mgr.14156) 130 : cluster [DBG] pgmap v93: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:35.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:35 vm02 bash[17011]: audit 2026-03-07T10:11:35.329211+0000 mon.a (mon.0) 318 : audit [DBG] from='client.? 192.168.123.102:0/336676369' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-07T10:11:35.711 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:35 vm02 bash[17011]: audit 2026-03-07T10:11:35.329211+0000 mon.a (mon.0) 318 : audit [DBG] from='client.? 192.168.123.102:0/336676369' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-07T10:11:37.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:37 vm10 bash[19283]: cluster 2026-03-07T10:11:36.286965+0000 mgr.a (mgr.14156) 131 : cluster [DBG] pgmap v94: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:37.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:37 vm10 bash[19283]: cluster 2026-03-07T10:11:36.286965+0000 mgr.a (mgr.14156) 131 : cluster [DBG] pgmap v94: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:37.961 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:37 vm02 bash[17011]: cluster 2026-03-07T10:11:36.286965+0000 mgr.a (mgr.14156) 131 : cluster [DBG] pgmap v94: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:37.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:37 vm02 bash[17011]: cluster 2026-03-07T10:11:36.286965+0000 mgr.a (mgr.14156) 131 : cluster [DBG] pgmap v94: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:39.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:39 vm10 bash[19283]: cluster 2026-03-07T10:11:38.287167+0000 mgr.a (mgr.14156) 132 : cluster [DBG] pgmap v95: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:39.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:39 vm10 bash[19283]: cluster 2026-03-07T10:11:38.287167+0000 mgr.a (mgr.14156) 132 : cluster [DBG] pgmap v95: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:39.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:39 vm02 bash[17011]: cluster 2026-03-07T10:11:38.287167+0000 mgr.a (mgr.14156) 132 : cluster [DBG] pgmap v95: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:39.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:39 vm02 bash[17011]: cluster 2026-03-07T10:11:38.287167+0000 mgr.a (mgr.14156) 132 : cluster [DBG] pgmap v95: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:40.219 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config 2026-03-07T10:11:40.619 INFO:teuthology.orchestra.run.vm02.stdout:Scheduled mon update... 2026-03-07T10:11:40.619 INFO:teuthology.orchestra.run.vm02.stdout:Scheduled container.foo update... 2026-03-07T10:11:40.619 INFO:teuthology.orchestra.run.vm02.stdout:Scheduled container.bar update... 2026-03-07T10:11:40.719 INFO:teuthology.run_tasks:Running task cephadm.wait_for_service... 2026-03-07T10:11:40.721 INFO:tasks.cephadm:Waiting for ceph service mon to start (timeout 300)... 2026-03-07T10:11:40.721 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph orch ls -f json 2026-03-07T10:11:41.495 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 07 10:11:41 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:11:41.495 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:11:41 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:11:41.495 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:41 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:11:41.756 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 07 10:11:41 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:11:41.756 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:11:41 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:11:41.756 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:41 vm02 bash[17011]: cluster 2026-03-07T10:11:40.287353+0000 mgr.a (mgr.14156) 133 : cluster [DBG] pgmap v96: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:41.756 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:41 vm02 bash[17011]: cluster 2026-03-07T10:11:40.287353+0000 mgr.a (mgr.14156) 133 : cluster [DBG] pgmap v96: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:41.756 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:41 vm02 bash[17011]: audit 2026-03-07T10:11:40.599500+0000 mgr.a (mgr.14156) 134 : audit [DBG] from='client.14280 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:11:41.756 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:41 vm02 bash[17011]: audit 2026-03-07T10:11:40.599500+0000 mgr.a (mgr.14156) 134 : audit [DBG] from='client.14280 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:11:41.756 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:41 vm02 bash[17011]: cephadm 2026-03-07T10:11:40.604525+0000 mgr.a (mgr.14156) 135 : cephadm [INF] Saving service mon spec with placement * 2026-03-07T10:11:41.756 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:41 vm02 bash[17011]: cephadm 2026-03-07T10:11:40.604525+0000 mgr.a (mgr.14156) 135 : cephadm [INF] Saving service mon spec with placement * 2026-03-07T10:11:41.756 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:41 vm02 bash[17011]: audit 2026-03-07T10:11:40.608880+0000 mon.a (mon.0) 319 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:41.756 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:41 vm02 bash[17011]: audit 2026-03-07T10:11:40.608880+0000 mon.a (mon.0) 319 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:41.756 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:41 vm02 bash[17011]: cephadm 2026-03-07T10:11:40.609658+0000 mgr.a (mgr.14156) 136 : cephadm [INF] Saving service container.foo spec with placement * 2026-03-07T10:11:41.756 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:41 vm02 bash[17011]: cephadm 2026-03-07T10:11:40.609658+0000 mgr.a (mgr.14156) 136 : cephadm [INF] Saving service container.foo spec with placement * 2026-03-07T10:11:41.756 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:41 vm02 bash[17011]: audit 2026-03-07T10:11:40.609857+0000 mon.a (mon.0) 320 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:11:41.756 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:41 vm02 bash[17011]: audit 2026-03-07T10:11:40.609857+0000 mon.a (mon.0) 320 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:11:41.756 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:41 vm02 bash[17011]: audit 2026-03-07T10:11:40.612903+0000 mon.a (mon.0) 321 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:41.756 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:41 vm02 bash[17011]: audit 2026-03-07T10:11:40.612903+0000 mon.a (mon.0) 321 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:41.756 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:41 vm02 bash[17011]: cephadm 2026-03-07T10:11:40.613672+0000 mgr.a (mgr.14156) 137 : cephadm [INF] Saving service container.bar spec with placement * 2026-03-07T10:11:41.756 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:41 vm02 bash[17011]: cephadm 2026-03-07T10:11:40.613672+0000 mgr.a (mgr.14156) 137 : cephadm [INF] Saving service container.bar spec with placement * 2026-03-07T10:11:41.756 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:41 vm02 bash[17011]: audit 2026-03-07T10:11:40.616939+0000 mon.a (mon.0) 322 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:41.756 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:41 vm02 bash[17011]: audit 2026-03-07T10:11:40.616939+0000 mon.a (mon.0) 322 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:41.756 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:41 vm02 bash[17011]: audit 2026-03-07T10:11:40.991630+0000 mon.a (mon.0) 323 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:11:41.756 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:41 vm02 bash[17011]: audit 2026-03-07T10:11:40.991630+0000 mon.a (mon.0) 323 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:11:41.756 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:41 vm02 bash[17011]: audit 2026-03-07T10:11:40.992540+0000 mon.a (mon.0) 324 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:11:41.756 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:41 vm02 bash[17011]: audit 2026-03-07T10:11:40.992540+0000 mon.a (mon.0) 324 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:11:41.756 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:41 vm02 bash[17011]: audit 2026-03-07T10:11:40.998669+0000 mon.a (mon.0) 325 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:41.756 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:41 vm02 bash[17011]: audit 2026-03-07T10:11:40.998669+0000 mon.a (mon.0) 325 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:41.756 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:41 vm02 bash[17011]: audit 2026-03-07T10:11:41.003967+0000 mon.a (mon.0) 326 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:41.756 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:41 vm02 bash[17011]: audit 2026-03-07T10:11:41.003967+0000 mon.a (mon.0) 326 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:41.756 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:41 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:11:41.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:41 vm10 bash[19283]: cluster 2026-03-07T10:11:40.287353+0000 mgr.a (mgr.14156) 133 : cluster [DBG] pgmap v96: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:41.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:41 vm10 bash[19283]: cluster 2026-03-07T10:11:40.287353+0000 mgr.a (mgr.14156) 133 : cluster [DBG] pgmap v96: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:41.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:41 vm10 bash[19283]: audit 2026-03-07T10:11:40.599500+0000 mgr.a (mgr.14156) 134 : audit [DBG] from='client.14280 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:11:41.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:41 vm10 bash[19283]: audit 2026-03-07T10:11:40.599500+0000 mgr.a (mgr.14156) 134 : audit [DBG] from='client.14280 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T10:11:41.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:41 vm10 bash[19283]: cephadm 2026-03-07T10:11:40.604525+0000 mgr.a (mgr.14156) 135 : cephadm [INF] Saving service mon spec with placement * 2026-03-07T10:11:41.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:41 vm10 bash[19283]: cephadm 2026-03-07T10:11:40.604525+0000 mgr.a (mgr.14156) 135 : cephadm [INF] Saving service mon spec with placement * 2026-03-07T10:11:41.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:41 vm10 bash[19283]: audit 2026-03-07T10:11:40.608880+0000 mon.a (mon.0) 319 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:41.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:41 vm10 bash[19283]: audit 2026-03-07T10:11:40.608880+0000 mon.a (mon.0) 319 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:41.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:41 vm10 bash[19283]: cephadm 2026-03-07T10:11:40.609658+0000 mgr.a (mgr.14156) 136 : cephadm [INF] Saving service container.foo spec with placement * 2026-03-07T10:11:41.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:41 vm10 bash[19283]: cephadm 2026-03-07T10:11:40.609658+0000 mgr.a (mgr.14156) 136 : cephadm [INF] Saving service container.foo spec with placement * 2026-03-07T10:11:41.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:41 vm10 bash[19283]: audit 2026-03-07T10:11:40.609857+0000 mon.a (mon.0) 320 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:11:41.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:41 vm10 bash[19283]: audit 2026-03-07T10:11:40.609857+0000 mon.a (mon.0) 320 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:11:41.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:41 vm10 bash[19283]: audit 2026-03-07T10:11:40.612903+0000 mon.a (mon.0) 321 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:41.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:41 vm10 bash[19283]: audit 2026-03-07T10:11:40.612903+0000 mon.a (mon.0) 321 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:41.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:41 vm10 bash[19283]: cephadm 2026-03-07T10:11:40.613672+0000 mgr.a (mgr.14156) 137 : cephadm [INF] Saving service container.bar spec with placement * 2026-03-07T10:11:41.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:41 vm10 bash[19283]: cephadm 2026-03-07T10:11:40.613672+0000 mgr.a (mgr.14156) 137 : cephadm [INF] Saving service container.bar spec with placement * 2026-03-07T10:11:41.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:41 vm10 bash[19283]: audit 2026-03-07T10:11:40.616939+0000 mon.a (mon.0) 322 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:41.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:41 vm10 bash[19283]: audit 2026-03-07T10:11:40.616939+0000 mon.a (mon.0) 322 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:41.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:41 vm10 bash[19283]: audit 2026-03-07T10:11:40.991630+0000 mon.a (mon.0) 323 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:11:41.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:41 vm10 bash[19283]: audit 2026-03-07T10:11:40.991630+0000 mon.a (mon.0) 323 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:11:41.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:41 vm10 bash[19283]: audit 2026-03-07T10:11:40.992540+0000 mon.a (mon.0) 324 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:11:41.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:41 vm10 bash[19283]: audit 2026-03-07T10:11:40.992540+0000 mon.a (mon.0) 324 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:11:41.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:41 vm10 bash[19283]: audit 2026-03-07T10:11:40.998669+0000 mon.a (mon.0) 325 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:41.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:41 vm10 bash[19283]: audit 2026-03-07T10:11:40.998669+0000 mon.a (mon.0) 325 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:41.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:41 vm10 bash[19283]: audit 2026-03-07T10:11:41.003967+0000 mon.a (mon.0) 326 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:41.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:41 vm10 bash[19283]: audit 2026-03-07T10:11:41.003967+0000 mon.a (mon.0) 326 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:42.339 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:42 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:11:42.339 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:11:42 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:11:42.339 INFO:journalctl@ceph.osd.1.vm10.stdout:Mar 07 10:11:42 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:11:42.695 INFO:journalctl@ceph.osd.1.vm10.stdout:Mar 07 10:11:42 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:11:42.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:42 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:11:42.695 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:11:42 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:11:43.039 INFO:journalctl@ceph.osd.1.vm10.stdout:Mar 07 10:11:42 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:11:43.039 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:42 vm10 bash[19283]: cephadm 2026-03-07T10:11:41.005640+0000 mgr.a (mgr.14156) 138 : cephadm [INF] Deploying daemon container.foo.vm02 on vm02 2026-03-07T10:11:43.039 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:42 vm10 bash[19283]: cephadm 2026-03-07T10:11:41.005640+0000 mgr.a (mgr.14156) 138 : cephadm [INF] Deploying daemon container.foo.vm02 on vm02 2026-03-07T10:11:43.039 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:42 vm10 bash[19283]: audit 2026-03-07T10:11:41.747268+0000 mon.a (mon.0) 327 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:43.039 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:42 vm10 bash[19283]: audit 2026-03-07T10:11:41.747268+0000 mon.a (mon.0) 327 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:43.039 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:42 vm10 bash[19283]: audit 2026-03-07T10:11:41.751959+0000 mon.a (mon.0) 328 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:43.039 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:42 vm10 bash[19283]: audit 2026-03-07T10:11:41.751959+0000 mon.a (mon.0) 328 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:43.039 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:42 vm10 bash[19283]: audit 2026-03-07T10:11:41.757673+0000 mon.a (mon.0) 329 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:43.039 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:42 vm10 bash[19283]: audit 2026-03-07T10:11:41.757673+0000 mon.a (mon.0) 329 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:43.039 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:42 vm10 bash[19283]: cephadm 2026-03-07T10:11:41.758588+0000 mgr.a (mgr.14156) 139 : cephadm [INF] Deploying daemon container.foo.vm10 on vm10 2026-03-07T10:11:43.039 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:42 vm10 bash[19283]: cephadm 2026-03-07T10:11:41.758588+0000 mgr.a (mgr.14156) 139 : cephadm [INF] Deploying daemon container.foo.vm10 on vm10 2026-03-07T10:11:43.039 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:42 vm10 bash[19283]: audit 2026-03-07T10:11:42.432948+0000 mon.a (mon.0) 330 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:43.039 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:42 vm10 bash[19283]: audit 2026-03-07T10:11:42.432948+0000 mon.a (mon.0) 330 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:43.039 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:42 vm10 bash[19283]: audit 2026-03-07T10:11:42.437455+0000 mon.a (mon.0) 331 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:43.039 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:42 vm10 bash[19283]: audit 2026-03-07T10:11:42.437455+0000 mon.a (mon.0) 331 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:43.039 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:42 vm10 bash[19283]: audit 2026-03-07T10:11:42.441629+0000 mon.a (mon.0) 332 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:43.039 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:42 vm10 bash[19283]: audit 2026-03-07T10:11:42.441629+0000 mon.a (mon.0) 332 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:43.039 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:42 vm10 bash[19283]: audit 2026-03-07T10:11:42.445260+0000 mon.a (mon.0) 333 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:43.039 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:42 vm10 bash[19283]: audit 2026-03-07T10:11:42.445260+0000 mon.a (mon.0) 333 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:43.039 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:42 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:11:43.039 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:11:42 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:11:43.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:42 vm02 bash[17011]: cephadm 2026-03-07T10:11:41.005640+0000 mgr.a (mgr.14156) 138 : cephadm [INF] Deploying daemon container.foo.vm02 on vm02 2026-03-07T10:11:43.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:42 vm02 bash[17011]: cephadm 2026-03-07T10:11:41.005640+0000 mgr.a (mgr.14156) 138 : cephadm [INF] Deploying daemon container.foo.vm02 on vm02 2026-03-07T10:11:43.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:42 vm02 bash[17011]: audit 2026-03-07T10:11:41.747268+0000 mon.a (mon.0) 327 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:43.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:42 vm02 bash[17011]: audit 2026-03-07T10:11:41.747268+0000 mon.a (mon.0) 327 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:43.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:42 vm02 bash[17011]: audit 2026-03-07T10:11:41.751959+0000 mon.a (mon.0) 328 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:43.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:42 vm02 bash[17011]: audit 2026-03-07T10:11:41.751959+0000 mon.a (mon.0) 328 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:43.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:42 vm02 bash[17011]: audit 2026-03-07T10:11:41.757673+0000 mon.a (mon.0) 329 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:43.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:42 vm02 bash[17011]: audit 2026-03-07T10:11:41.757673+0000 mon.a (mon.0) 329 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:43.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:42 vm02 bash[17011]: cephadm 2026-03-07T10:11:41.758588+0000 mgr.a (mgr.14156) 139 : cephadm [INF] Deploying daemon container.foo.vm10 on vm10 2026-03-07T10:11:43.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:42 vm02 bash[17011]: cephadm 2026-03-07T10:11:41.758588+0000 mgr.a (mgr.14156) 139 : cephadm [INF] Deploying daemon container.foo.vm10 on vm10 2026-03-07T10:11:43.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:42 vm02 bash[17011]: audit 2026-03-07T10:11:42.432948+0000 mon.a (mon.0) 330 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:43.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:42 vm02 bash[17011]: audit 2026-03-07T10:11:42.432948+0000 mon.a (mon.0) 330 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:43.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:42 vm02 bash[17011]: audit 2026-03-07T10:11:42.437455+0000 mon.a (mon.0) 331 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:43.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:42 vm02 bash[17011]: audit 2026-03-07T10:11:42.437455+0000 mon.a (mon.0) 331 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:43.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:42 vm02 bash[17011]: audit 2026-03-07T10:11:42.441629+0000 mon.a (mon.0) 332 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:43.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:42 vm02 bash[17011]: audit 2026-03-07T10:11:42.441629+0000 mon.a (mon.0) 332 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:43.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:42 vm02 bash[17011]: audit 2026-03-07T10:11:42.445260+0000 mon.a (mon.0) 333 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:43.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:42 vm02 bash[17011]: audit 2026-03-07T10:11:42.445260+0000 mon.a (mon.0) 333 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:43.445 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:43 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:11:43.445 INFO:journalctl@ceph.osd.1.vm10.stdout:Mar 07 10:11:43 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:11:43.446 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:11:43 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:11:43.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:43 vm02 bash[17011]: cluster 2026-03-07T10:11:42.287582+0000 mgr.a (mgr.14156) 140 : cluster [DBG] pgmap v97: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:43.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:43 vm02 bash[17011]: cluster 2026-03-07T10:11:42.287582+0000 mgr.a (mgr.14156) 140 : cluster [DBG] pgmap v97: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:43.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:43 vm02 bash[17011]: cephadm 2026-03-07T10:11:42.446573+0000 mgr.a (mgr.14156) 141 : cephadm [INF] Deploying daemon container.bar.vm10 on vm10 2026-03-07T10:11:43.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:43 vm02 bash[17011]: cephadm 2026-03-07T10:11:42.446573+0000 mgr.a (mgr.14156) 141 : cephadm [INF] Deploying daemon container.bar.vm10 on vm10 2026-03-07T10:11:44.191 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:43 vm10 bash[19283]: cluster 2026-03-07T10:11:42.287582+0000 mgr.a (mgr.14156) 140 : cluster [DBG] pgmap v97: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:44.191 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:43 vm10 bash[19283]: cluster 2026-03-07T10:11:42.287582+0000 mgr.a (mgr.14156) 140 : cluster [DBG] pgmap v97: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:44.191 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:43 vm10 bash[19283]: cephadm 2026-03-07T10:11:42.446573+0000 mgr.a (mgr.14156) 141 : cephadm [INF] Deploying daemon container.bar.vm10 on vm10 2026-03-07T10:11:44.191 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:43 vm10 bash[19283]: cephadm 2026-03-07T10:11:42.446573+0000 mgr.a (mgr.14156) 141 : cephadm [INF] Deploying daemon container.bar.vm10 on vm10 2026-03-07T10:11:45.488 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config 2026-03-07T10:11:45.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:45 vm10 bash[19283]: cluster 2026-03-07T10:11:44.287828+0000 mgr.a (mgr.14156) 142 : cluster [DBG] pgmap v98: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:45.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:45 vm10 bash[19283]: cluster 2026-03-07T10:11:44.287828+0000 mgr.a (mgr.14156) 142 : cluster [DBG] pgmap v98: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:45.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:45 vm10 bash[19283]: audit 2026-03-07T10:11:44.351553+0000 mon.a (mon.0) 334 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:45.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:45 vm10 bash[19283]: audit 2026-03-07T10:11:44.351553+0000 mon.a (mon.0) 334 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:45.706 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:45 vm02 bash[17011]: cluster 2026-03-07T10:11:44.287828+0000 mgr.a (mgr.14156) 142 : cluster [DBG] pgmap v98: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:45.706 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:45 vm02 bash[17011]: cluster 2026-03-07T10:11:44.287828+0000 mgr.a (mgr.14156) 142 : cluster [DBG] pgmap v98: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:45.706 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:45 vm02 bash[17011]: audit 2026-03-07T10:11:44.351553+0000 mon.a (mon.0) 334 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:45.706 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:45 vm02 bash[17011]: audit 2026-03-07T10:11:44.351553+0000 mon.a (mon.0) 334 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:46.091 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-07T10:11:46.091 INFO:teuthology.orchestra.run.vm02.stdout:[{"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-07T10:11:40.617658Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-07T10:11:40.613675Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-07T10:11:42.445590Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-07T10:11:40.609750Z", "running": 0, "size": 2}}, {"events": ["2026-03-07T10:09:30.906338Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm02=a", "vm10=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-07T10:09:30.085053Z", "last_refresh": "2026-03-07T10:10:05.919008Z", "running": 2, "size": 2}}, {"events": ["2026-03-07T10:11:41.004615Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-07T10:11:40.604541Z", "last_refresh": "2026-03-07T10:10:05.919095Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "8bccc98d839aa18345ec1336292d0452ca331737e49f12524f635044dcabcfe1", "container_image_name": "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:ffa52c72fad7bdd2657408de9cf8d87fc2c72f716d1a00277ba13f7c12b404e0", "last_refresh": "2026-03-07T10:10:05.919059Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-07T10:11:46.278 INFO:tasks.cephadm:mon has 2/2 2026-03-07T10:11:46.278 INFO:teuthology.run_tasks:Running task cephadm.wait_for_service... 2026-03-07T10:11:46.281 INFO:tasks.cephadm:Waiting for ceph service container.foo to start (timeout 300)... 2026-03-07T10:11:46.281 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph orch ls -f json 2026-03-07T10:11:47.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:47 vm02 bash[17011]: audit 2026-03-07T10:11:46.081451+0000 mgr.a (mgr.14156) 143 : audit [DBG] from='client.14284 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T10:11:47.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:47 vm02 bash[17011]: audit 2026-03-07T10:11:46.081451+0000 mgr.a (mgr.14156) 143 : audit [DBG] from='client.14284 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T10:11:47.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:47 vm02 bash[17011]: cluster 2026-03-07T10:11:46.288107+0000 mgr.a (mgr.14156) 144 : cluster [DBG] pgmap v99: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:47.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:47 vm02 bash[17011]: cluster 2026-03-07T10:11:46.288107+0000 mgr.a (mgr.14156) 144 : cluster [DBG] pgmap v99: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:47.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:47 vm10 bash[19283]: audit 2026-03-07T10:11:46.081451+0000 mgr.a (mgr.14156) 143 : audit [DBG] from='client.14284 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T10:11:47.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:47 vm10 bash[19283]: audit 2026-03-07T10:11:46.081451+0000 mgr.a (mgr.14156) 143 : audit [DBG] from='client.14284 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T10:11:47.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:47 vm10 bash[19283]: cluster 2026-03-07T10:11:46.288107+0000 mgr.a (mgr.14156) 144 : cluster [DBG] pgmap v99: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:47.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:47 vm10 bash[19283]: cluster 2026-03-07T10:11:46.288107+0000 mgr.a (mgr.14156) 144 : cluster [DBG] pgmap v99: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:51.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:51 vm02 bash[17011]: cluster 2026-03-07T10:11:48.288392+0000 mgr.a (mgr.14156) 145 : cluster [DBG] pgmap v100: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:51.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:51 vm02 bash[17011]: cluster 2026-03-07T10:11:48.288392+0000 mgr.a (mgr.14156) 145 : cluster [DBG] pgmap v100: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:51.905 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:51 vm10 bash[19283]: cluster 2026-03-07T10:11:48.288392+0000 mgr.a (mgr.14156) 145 : cluster [DBG] pgmap v100: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:51.905 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:51 vm10 bash[19283]: cluster 2026-03-07T10:11:48.288392+0000 mgr.a (mgr.14156) 145 : cluster [DBG] pgmap v100: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:52.053 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config 2026-03-07T10:11:52.837 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:52 vm10 bash[19283]: cluster 2026-03-07T10:11:50.288634+0000 mgr.a (mgr.14156) 146 : cluster [DBG] pgmap v101: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:52.838 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:52 vm10 bash[19283]: cluster 2026-03-07T10:11:50.288634+0000 mgr.a (mgr.14156) 146 : cluster [DBG] pgmap v101: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:52.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:52 vm02 bash[17011]: cluster 2026-03-07T10:11:50.288634+0000 mgr.a (mgr.14156) 146 : cluster [DBG] pgmap v101: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:52.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:52 vm02 bash[17011]: cluster 2026-03-07T10:11:50.288634+0000 mgr.a (mgr.14156) 146 : cluster [DBG] pgmap v101: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:53.012 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-07T10:11:53.012 INFO:teuthology.orchestra.run.vm02.stdout:[{"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-07T10:11:40.617658Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-07T10:11:40.613675Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-07T10:11:42.445590Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-07T10:11:40.609750Z", "running": 0, "size": 2}}, {"events": ["2026-03-07T10:09:30.906338Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm02=a", "vm10=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-07T10:09:30.085053Z", "last_refresh": "2026-03-07T10:10:05.919008Z", "running": 2, "size": 2}}, {"events": ["2026-03-07T10:11:41.004615Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-07T10:11:40.604541Z", "last_refresh": "2026-03-07T10:10:05.919095Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "8bccc98d839aa18345ec1336292d0452ca331737e49f12524f635044dcabcfe1", "container_image_name": "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:ffa52c72fad7bdd2657408de9cf8d87fc2c72f716d1a00277ba13f7c12b404e0", "last_refresh": "2026-03-07T10:10:05.919059Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-07T10:11:53.073 INFO:tasks.cephadm:container.foo has 0/2 2026-03-07T10:11:53.712 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 07 10:11:53 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:11:53.712 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:11:53 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:11:53.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:53 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:11:53.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:53 vm02 bash[17011]: cluster 2026-03-07T10:11:52.288857+0000 mgr.a (mgr.14156) 147 : cluster [DBG] pgmap v102: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:53.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:53 vm02 bash[17011]: cluster 2026-03-07T10:11:52.288857+0000 mgr.a (mgr.14156) 147 : cluster [DBG] pgmap v102: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:53.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:53 vm02 bash[17011]: audit 2026-03-07T10:11:53.170885+0000 mon.a (mon.0) 335 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:53.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:53 vm02 bash[17011]: audit 2026-03-07T10:11:53.170885+0000 mon.a (mon.0) 335 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:53.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:53 vm02 bash[17011]: audit 2026-03-07T10:11:53.176203+0000 mon.a (mon.0) 336 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:53.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:53 vm02 bash[17011]: audit 2026-03-07T10:11:53.176203+0000 mon.a (mon.0) 336 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:53.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:53 vm02 bash[17011]: audit 2026-03-07T10:11:53.181191+0000 mon.a (mon.0) 337 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:53.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:53 vm02 bash[17011]: audit 2026-03-07T10:11:53.181191+0000 mon.a (mon.0) 337 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:53.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:53 vm10 bash[19283]: cluster 2026-03-07T10:11:52.288857+0000 mgr.a (mgr.14156) 147 : cluster [DBG] pgmap v102: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:53.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:53 vm10 bash[19283]: cluster 2026-03-07T10:11:52.288857+0000 mgr.a (mgr.14156) 147 : cluster [DBG] pgmap v102: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:53.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:53 vm10 bash[19283]: audit 2026-03-07T10:11:53.170885+0000 mon.a (mon.0) 335 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:53.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:53 vm10 bash[19283]: audit 2026-03-07T10:11:53.170885+0000 mon.a (mon.0) 335 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:53.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:53 vm10 bash[19283]: audit 2026-03-07T10:11:53.176203+0000 mon.a (mon.0) 336 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:53.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:53 vm10 bash[19283]: audit 2026-03-07T10:11:53.176203+0000 mon.a (mon.0) 336 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:53.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:53 vm10 bash[19283]: audit 2026-03-07T10:11:53.181191+0000 mon.a (mon.0) 337 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:53.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:53 vm10 bash[19283]: audit 2026-03-07T10:11:53.181191+0000 mon.a (mon.0) 337 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:11:54.074 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph orch ls -f json 2026-03-07T10:11:54.081 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:53 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:11:54.081 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:11:53 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:11:54.081 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 07 10:11:53 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:11:54.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:54 vm10 bash[19283]: audit 2026-03-07T10:11:53.010280+0000 mgr.a (mgr.14156) 148 : audit [DBG] from='client.14288 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T10:11:54.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:54 vm10 bash[19283]: audit 2026-03-07T10:11:53.010280+0000 mgr.a (mgr.14156) 148 : audit [DBG] from='client.14288 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T10:11:54.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:54 vm10 bash[19283]: cephadm 2026-03-07T10:11:53.181940+0000 mgr.a (mgr.14156) 149 : cephadm [INF] Deploying daemon container.bar.vm02 on vm02 2026-03-07T10:11:54.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:54 vm10 bash[19283]: cephadm 2026-03-07T10:11:53.181940+0000 mgr.a (mgr.14156) 149 : cephadm [INF] Deploying daemon container.bar.vm02 on vm02 2026-03-07T10:11:54.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:54 vm02 bash[17011]: audit 2026-03-07T10:11:53.010280+0000 mgr.a (mgr.14156) 148 : audit [DBG] from='client.14288 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T10:11:54.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:54 vm02 bash[17011]: audit 2026-03-07T10:11:53.010280+0000 mgr.a (mgr.14156) 148 : audit [DBG] from='client.14288 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T10:11:54.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:54 vm02 bash[17011]: cephadm 2026-03-07T10:11:53.181940+0000 mgr.a (mgr.14156) 149 : cephadm [INF] Deploying daemon container.bar.vm02 on vm02 2026-03-07T10:11:54.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:54 vm02 bash[17011]: cephadm 2026-03-07T10:11:53.181940+0000 mgr.a (mgr.14156) 149 : cephadm [INF] Deploying daemon container.bar.vm02 on vm02 2026-03-07T10:11:55.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:55 vm10 bash[19283]: cluster 2026-03-07T10:11:54.289091+0000 mgr.a (mgr.14156) 150 : cluster [DBG] pgmap v103: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:55.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:55 vm10 bash[19283]: cluster 2026-03-07T10:11:54.289091+0000 mgr.a (mgr.14156) 150 : cluster [DBG] pgmap v103: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:55.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:55 vm02 bash[17011]: cluster 2026-03-07T10:11:54.289091+0000 mgr.a (mgr.14156) 150 : cluster [DBG] pgmap v103: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:55.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:55 vm02 bash[17011]: cluster 2026-03-07T10:11:54.289091+0000 mgr.a (mgr.14156) 150 : cluster [DBG] pgmap v103: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:57.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:57 vm02 bash[17011]: cluster 2026-03-07T10:11:56.289313+0000 mgr.a (mgr.14156) 151 : cluster [DBG] pgmap v104: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:57.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:57 vm02 bash[17011]: cluster 2026-03-07T10:11:56.289313+0000 mgr.a (mgr.14156) 151 : cluster [DBG] pgmap v104: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:57.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:57 vm10 bash[19283]: cluster 2026-03-07T10:11:56.289313+0000 mgr.a (mgr.14156) 151 : cluster [DBG] pgmap v104: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:57.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:57 vm10 bash[19283]: cluster 2026-03-07T10:11:56.289313+0000 mgr.a (mgr.14156) 151 : cluster [DBG] pgmap v104: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:58.824 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config 2026-03-07T10:11:59.916 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:59 vm02 bash[17011]: cluster 2026-03-07T10:11:58.289542+0000 mgr.a (mgr.14156) 152 : cluster [DBG] pgmap v105: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:11:59.916 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:11:59 vm02 bash[17011]: cluster 2026-03-07T10:11:58.289542+0000 mgr.a (mgr.14156) 152 : cluster [DBG] pgmap v105: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:00.145 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-07T10:12:00.145 INFO:teuthology.orchestra.run.vm02.stdout:[{"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-07T10:11:53.181419Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-07T10:11:40.613675Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-07T10:11:42.445590Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-07T10:11:40.609750Z", "running": 0, "size": 2}}, {"events": ["2026-03-07T10:09:30.906338Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm02=a", "vm10=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-07T10:09:30.085053Z", "last_refresh": "2026-03-07T10:10:05.919008Z", "running": 2, "size": 2}}, {"events": ["2026-03-07T10:11:41.004615Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-07T10:11:40.604541Z", "last_refresh": "2026-03-07T10:10:05.919095Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "8bccc98d839aa18345ec1336292d0452ca331737e49f12524f635044dcabcfe1", "container_image_name": "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:ffa52c72fad7bdd2657408de9cf8d87fc2c72f716d1a00277ba13f7c12b404e0", "last_refresh": "2026-03-07T10:10:05.919059Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-07T10:12:00.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:59 vm10 bash[19283]: cluster 2026-03-07T10:11:58.289542+0000 mgr.a (mgr.14156) 152 : cluster [DBG] pgmap v105: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:00.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:11:59 vm10 bash[19283]: cluster 2026-03-07T10:11:58.289542+0000 mgr.a (mgr.14156) 152 : cluster [DBG] pgmap v105: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:00.233 INFO:tasks.cephadm:container.foo has 0/2 2026-03-07T10:12:01.234 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph orch ls -f json 2026-03-07T10:12:01.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:01 vm02 bash[17011]: audit 2026-03-07T10:12:00.141931+0000 mgr.a (mgr.14156) 153 : audit [DBG] from='client.24181 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T10:12:01.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:01 vm02 bash[17011]: audit 2026-03-07T10:12:00.141931+0000 mgr.a (mgr.14156) 153 : audit [DBG] from='client.24181 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T10:12:01.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:01 vm02 bash[17011]: cluster 2026-03-07T10:12:00.289776+0000 mgr.a (mgr.14156) 154 : cluster [DBG] pgmap v106: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:01.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:01 vm02 bash[17011]: cluster 2026-03-07T10:12:00.289776+0000 mgr.a (mgr.14156) 154 : cluster [DBG] pgmap v106: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:01.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:01 vm02 bash[17011]: audit 2026-03-07T10:12:00.503732+0000 mon.a (mon.0) 338 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:01.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:01 vm02 bash[17011]: audit 2026-03-07T10:12:00.503732+0000 mon.a (mon.0) 338 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:01.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:01 vm02 bash[17011]: audit 2026-03-07T10:12:00.507638+0000 mon.a (mon.0) 339 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:01.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:01 vm02 bash[17011]: audit 2026-03-07T10:12:00.507638+0000 mon.a (mon.0) 339 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:01.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:01 vm02 bash[17011]: audit 2026-03-07T10:12:00.512325+0000 mon.a (mon.0) 340 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:01.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:01 vm02 bash[17011]: audit 2026-03-07T10:12:00.512325+0000 mon.a (mon.0) 340 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:01.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:01 vm02 bash[17011]: audit 2026-03-07T10:12:00.517297+0000 mon.a (mon.0) 341 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:01.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:01 vm02 bash[17011]: audit 2026-03-07T10:12:00.517297+0000 mon.a (mon.0) 341 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:01.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:01 vm02 bash[17011]: audit 2026-03-07T10:12:00.532428+0000 mon.a (mon.0) 342 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:12:01.801 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:01 vm02 bash[17011]: audit 2026-03-07T10:12:00.532428+0000 mon.a (mon.0) 342 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:12:01.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:01 vm10 bash[19283]: audit 2026-03-07T10:12:00.141931+0000 mgr.a (mgr.14156) 153 : audit [DBG] from='client.24181 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T10:12:01.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:01 vm10 bash[19283]: audit 2026-03-07T10:12:00.141931+0000 mgr.a (mgr.14156) 153 : audit [DBG] from='client.24181 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T10:12:01.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:01 vm10 bash[19283]: cluster 2026-03-07T10:12:00.289776+0000 mgr.a (mgr.14156) 154 : cluster [DBG] pgmap v106: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:01.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:01 vm10 bash[19283]: cluster 2026-03-07T10:12:00.289776+0000 mgr.a (mgr.14156) 154 : cluster [DBG] pgmap v106: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:01.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:01 vm10 bash[19283]: audit 2026-03-07T10:12:00.503732+0000 mon.a (mon.0) 338 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:01.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:01 vm10 bash[19283]: audit 2026-03-07T10:12:00.503732+0000 mon.a (mon.0) 338 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:01.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:01 vm10 bash[19283]: audit 2026-03-07T10:12:00.507638+0000 mon.a (mon.0) 339 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:01.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:01 vm10 bash[19283]: audit 2026-03-07T10:12:00.507638+0000 mon.a (mon.0) 339 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:01.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:01 vm10 bash[19283]: audit 2026-03-07T10:12:00.512325+0000 mon.a (mon.0) 340 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:01.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:01 vm10 bash[19283]: audit 2026-03-07T10:12:00.512325+0000 mon.a (mon.0) 340 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:01.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:01 vm10 bash[19283]: audit 2026-03-07T10:12:00.517297+0000 mon.a (mon.0) 341 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:01.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:01 vm10 bash[19283]: audit 2026-03-07T10:12:00.517297+0000 mon.a (mon.0) 341 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:01.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:01 vm10 bash[19283]: audit 2026-03-07T10:12:00.532428+0000 mon.a (mon.0) 342 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:12:01.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:01 vm10 bash[19283]: audit 2026-03-07T10:12:00.532428+0000 mon.a (mon.0) 342 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:12:03.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:03 vm10 bash[19283]: cluster 2026-03-07T10:12:02.290090+0000 mgr.a (mgr.14156) 155 : cluster [DBG] pgmap v107: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:03.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:03 vm10 bash[19283]: cluster 2026-03-07T10:12:02.290090+0000 mgr.a (mgr.14156) 155 : cluster [DBG] pgmap v107: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:03.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:03 vm02 bash[17011]: cluster 2026-03-07T10:12:02.290090+0000 mgr.a (mgr.14156) 155 : cluster [DBG] pgmap v107: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:03.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:03 vm02 bash[17011]: cluster 2026-03-07T10:12:02.290090+0000 mgr.a (mgr.14156) 155 : cluster [DBG] pgmap v107: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:05.313 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config 2026-03-07T10:12:05.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:05 vm10 bash[19283]: cluster 2026-03-07T10:12:04.290367+0000 mgr.a (mgr.14156) 156 : cluster [DBG] pgmap v108: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:05.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:05 vm10 bash[19283]: cluster 2026-03-07T10:12:04.290367+0000 mgr.a (mgr.14156) 156 : cluster [DBG] pgmap v108: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:05.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:05 vm10 bash[19283]: audit 2026-03-07T10:12:04.356556+0000 mon.a (mon.0) 343 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:05.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:05 vm10 bash[19283]: audit 2026-03-07T10:12:04.356556+0000 mon.a (mon.0) 343 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:05.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:05 vm02 bash[17011]: cluster 2026-03-07T10:12:04.290367+0000 mgr.a (mgr.14156) 156 : cluster [DBG] pgmap v108: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:05.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:05 vm02 bash[17011]: cluster 2026-03-07T10:12:04.290367+0000 mgr.a (mgr.14156) 156 : cluster [DBG] pgmap v108: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:05.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:05 vm02 bash[17011]: audit 2026-03-07T10:12:04.356556+0000 mon.a (mon.0) 343 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:05.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:05 vm02 bash[17011]: audit 2026-03-07T10:12:04.356556+0000 mon.a (mon.0) 343 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:05.765 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-07T10:12:05.765 INFO:teuthology.orchestra.run.vm02.stdout:[{"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-07T10:12:00.517521Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-07T10:11:40.613675Z", "last_refresh": "2026-03-07T10:12:05.485847Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-07T10:11:42.445590Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-07T10:11:40.609750Z", "last_refresh": "2026-03-07T10:12:05.485681Z", "running": 2, "size": 2}}, {"events": ["2026-03-07T10:09:30.906338Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm02=a", "vm10=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-07T10:09:30.085053Z", "last_refresh": "2026-03-07T10:12:05.485817Z", "running": 2, "size": 2}}, {"events": ["2026-03-07T10:11:41.004615Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-07T10:11:40.604541Z", "last_refresh": "2026-03-07T10:12:05.485750Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "8bccc98d839aa18345ec1336292d0452ca331737e49f12524f635044dcabcfe1", "container_image_name": "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:ffa52c72fad7bdd2657408de9cf8d87fc2c72f716d1a00277ba13f7c12b404e0", "last_refresh": "2026-03-07T10:12:05.485786Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-07T10:12:05.846 INFO:tasks.cephadm:container.foo has 2/2 2026-03-07T10:12:05.846 INFO:teuthology.run_tasks:Running task cephadm.wait_for_service... 2026-03-07T10:12:05.848 INFO:tasks.cephadm:Waiting for ceph service container.bar to start (timeout 300)... 2026-03-07T10:12:05.848 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e -- ceph orch ls -f json 2026-03-07T10:12:06.462 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 07 10:12:06 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:12:06.462 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:12:06 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:12:06.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:12:06.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 systemd[1]: Stopping Ceph mon.a for 6c715b7a-1a0d-11f1-b180-89615ccd948e... 2026-03-07T10:12:06.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[17011]: debug 2026-03-07T10:12:06.261+0000 7f5cf8f36640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-07T10:12:06.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[17011]: debug 2026-03-07T10:12:06.261+0000 7f5cf8f36640 -1 mon.a@0(leader) e2 *** Got Signal Terminated *** 2026-03-07T10:12:06.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34174]: ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e-mon-a 2026-03-07T10:12:06.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 systemd[1]: ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@mon.a.service: Deactivated successfully. 2026-03-07T10:12:06.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 systemd[1]: Stopped Ceph mon.a for 6c715b7a-1a0d-11f1-b180-89615ccd948e. 2026-03-07T10:12:06.962 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 07 10:12:06 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:12:06.962 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:12:06 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:12:06.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:12:06.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 systemd[1]: Started Ceph mon.a for 6c715b7a-1a0d-11f1-b180-89615ccd948e. 2026-03-07T10:12:06.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.621+0000 7f9f9dc93d80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-07T10:12:06.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.621+0000 7f9f9dc93d80 0 ceph version 19.2.3-39-g340d3c24fc6 (340d3c24fc6ae7529322dc7ccee6c6cb2589da0a) squid (stable), process ceph-mon, pid 7 2026-03-07T10:12:06.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.621+0000 7f9f9dc93d80 0 pidfile_write: ignore empty --pid-file 2026-03-07T10:12:06.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 0 load: jerasure load: lrc 2026-03-07T10:12:06.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: RocksDB version: 7.9.2 2026-03-07T10:12:06.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Git sha 0 2026-03-07T10:12:06.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Compile date 2026-03-06 13:52:12 2026-03-07T10:12:06.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: DB SUMMARY 2026-03-07T10:12:06.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: DB Session ID: GGGFPN3ZWMWY5LHRCL4U 2026-03-07T10:12:06.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: CURRENT file: CURRENT 2026-03-07T10:12:06.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-07T10:12:06.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: MANIFEST file: MANIFEST-000015 size: 281 Bytes 2026-03-07T10:12:06.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 2, files: 000008.sst 000013.sst 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 000014.log size: 3902767 ; 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.error_if_exists: 0 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.create_if_missing: 0 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.paranoid_checks: 1 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.env: 0x55f1513ebca0 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.info_log: 0x55f17b2ef700 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.statistics: (nil) 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.use_fsync: 0 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.max_log_file_size: 0 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.allow_fallocate: 1 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.use_direct_reads: 0 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.db_log_dir: 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.wal_dir: 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.write_buffer_manager: 0x55f17b2f3900 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-07T10:12:06.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-07T10:12:06.964 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-07T10:12:06.964 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-07T10:12:06.964 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-07T10:12:06.964 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.unordered_write: 0 2026-03-07T10:12:06.964 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-07T10:12:06.964 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-07T10:12:06.964 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-07T10:12:06.964 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-07T10:12:06.964 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.row_cache: None 2026-03-07T10:12:06.964 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.wal_filter: None 2026-03-07T10:12:06.964 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-07T10:12:06.964 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-07T10:12:06.964 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.two_write_queues: 0 2026-03-07T10:12:06.964 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-07T10:12:06.964 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.wal_compression: 0 2026-03-07T10:12:06.964 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.atomic_flush: 0 2026-03-07T10:12:06.964 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-07T10:12:06.964 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-07T10:12:06.964 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-07T10:12:06.964 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.log_readahead_size: 0 2026-03-07T10:12:06.964 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-07T10:12:06.964 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-07T10:12:06.964 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-07T10:12:06.964 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.max_background_jobs: 2 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.max_background_compactions: -1 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.max_subcompactions: 1 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.max_open_files: -1 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.max_background_flushes: -1 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Compression algorithms supported: 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: kZSTD supported: 0 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: kXpressCompression supported: 0 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: kBZip2Compression supported: 0 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: kLZ4Compression supported: 1 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: kZlibCompression supported: 1 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: kSnappyCompression supported: 1 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000015 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.merge_operator: 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.compaction_filter: None 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55f17b2ee080) 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: cache_index_and_filter_blocks: 1 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: pin_top_level_index_and_filter: 1 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: index_type: 0 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: data_block_index_type: 0 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: index_shortening: 1 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: data_block_hash_table_util_ratio: 0.750000 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: checksum: 4 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: no_block_cache: 0 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: block_cache: 0x55f17b3151f0 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: block_cache_name: BinnedLRUCache 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: block_cache_options: 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: capacity : 536870912 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: num_shard_bits : 4 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: strict_capacity_limit : 0 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: high_pri_pool_ratio: 0.000 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: block_cache_compressed: (nil) 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: persistent_cache: (nil) 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: block_size: 4096 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: block_size_deviation: 10 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: block_restart_interval: 16 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: index_block_restart_interval: 1 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: metadata_block_size: 4096 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: partition_filters: 0 2026-03-07T10:12:06.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: use_delta_encoding: 1 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: filter_policy: bloomfilter 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: whole_key_filtering: 1 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: verify_compression: 0 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: read_amp_bytes_per_bit: 0 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: format_version: 5 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: enable_index_compression: 1 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: block_align: 0 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: max_auto_readahead_size: 262144 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: prepopulate_block_cache: 0 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: initial_auto_readahead_size: 8192 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: num_file_reads_for_auto_readahead: 2 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.compression: NoCompression 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.num_levels: 7 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-07T10:12:06.966 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.inplace_update_support: 0 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.bloom_locality: 0 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.max_successive_merges: 0 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.ttl: 2592000 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.enable_blob_files: false 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.min_blob_size: 0 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.625+0000 7f9f9dc93d80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.633+0000 7f9f9dc93d80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000015 succeeded,manifest_file_number is 15, next_file_number is 17, last_sequence is 225, log_number is 10,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 10 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.633+0000 7f9f9dc93d80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 10 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.633+0000 7f9f9dc93d80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: ab4b5d8f-62d9-446c-92eb-b6ba735315c3 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.633+0000 7f9f9dc93d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1772878326636475, "job": 1, "event": "recovery_started", "wal_files": [14]} 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.633+0000 7f9f9dc93d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #14 mode 2 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.649+0000 7f9f9dc93d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1772878326652140, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 18, "file_size": 3438273, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 230, "largest_seqno": 3499, "table_properties": {"data_size": 3426949, "index_size": 6699, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 3589, "raw_key_size": 31169, "raw_average_key_size": 22, "raw_value_size": 3400126, "raw_average_value_size": 2402, "num_data_blocks": 317, "num_entries": 1415, "num_filter_entries": 1415, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1772878326, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "ab4b5d8f-62d9-446c-92eb-b6ba735315c3", "db_session_id": "GGGFPN3ZWMWY5LHRCL4U", "orig_file_number": 18, "seqno_to_time_mapping": "N/A"}} 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.649+0000 7f9f9dc93d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1772878326652261, "job": 1, "event": "recovery_finished"} 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.649+0000 7f9f9dc93d80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 20 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.653+0000 7f9f9dc93d80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-a/store.db/000014.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.653+0000 7f9f9dc93d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55f17b316e00 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.653+0000 7f9f9dc93d80 4 rocksdb: DB pointer 0x55f17b426000 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.653+0000 7f9f9dc93d80 0 starting mon.a rank 0 at public addrs [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] at bind addrs [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon_data /var/lib/ceph/mon/ceph-a fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.653+0000 7f9f9dc93d80 1 mon.a@-1(???) e2 preinit fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.653+0000 7f9f9dc93d80 0 mon.a@-1(???).mds e1 new map 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.653+0000 7f9f9dc93d80 0 mon.a@-1(???).mds e1 print_map 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: e1 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: btime 2026-03-07T10:07:53:157919+0000 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: enable_multiple, ever_enabled_multiple: 1,1 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: legacy client fscid: -1 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: No filesystems configured 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.653+0000 7f9f9dc93d80 0 mon.a@-1(???).osd e14 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.653+0000 7f9f9dc93d80 0 mon.a@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.653+0000 7f9f9dc93d80 0 mon.a@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.653+0000 7f9f9dc93d80 0 mon.a@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-07T10:12:06.967 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:06 vm02 bash[34286]: debug 2026-03-07T10:12:06.653+0000 7f9f9dc93d80 1 mon.a@-1(???).paxosservice(auth 1..7) refresh upgraded, format 0 -> 3 2026-03-07T10:12:07.581 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: audit 2026-03-07T10:12:06.272624+0000 mon.b (mon.1) 10 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-07T10:12:07.581 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: audit 2026-03-07T10:12:06.272624+0000 mon.b (mon.1) 10 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-07T10:12:07.581 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: audit 2026-03-07T10:12:06.272972+0000 mon.b (mon.1) 11 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:12:07.581 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: audit 2026-03-07T10:12:06.272972+0000 mon.b (mon.1) 11 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:12:07.581 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:06.290579+0000 mgr.a (mgr.14156) 160 : cluster [DBG] pgmap v109: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:07.581 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:06.290579+0000 mgr.a (mgr.14156) 160 : cluster [DBG] pgmap v109: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:07.581 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:07.266454+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-07T10:12:07.581 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:07.266454+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-07T10:12:07.581 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:07.269253+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-07T10:12:07.581 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:07.269253+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-07T10:12:07.581 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:07.273462+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-07T10:12:07.581 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:07.273462+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-07T10:12:07.581 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:07.273471+0000 mon.a (mon.0) 4 : cluster [DBG] fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:12:07.581 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:07.273471+0000 mon.a (mon.0) 4 : cluster [DBG] fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:07.273476+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-07T10:09:14.324495+0000 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:07.273476+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-07T10:09:14.324495+0000 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:07.273480+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-07T10:07:51.937005+0000 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:07.273480+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-07T10:07:51.937005+0000 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:07.273484+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 19 (squid) 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:07.273484+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 19 (squid) 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:07.273489+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:07.273489+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:07.273493+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:07.273493+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:07.273498+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon.b 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:07.273498+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon.b 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:07.273870+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:07.273870+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:07.273893+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:07.273893+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:07.274579+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e14: a(active, since 3m), standbys: b 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:07.274579+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e14: a(active, since 3m), standbys: b 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:07.274687+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: cluster 2026-03-07T10:12:07.274687+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: audit 2026-03-07T10:12:07.283586+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14156 ' entity='' 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: audit 2026-03-07T10:12:07.283586+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14156 ' entity='' 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: audit 2026-03-07T10:12:07.292487+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14156 ' entity='mgr.a' 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: audit 2026-03-07T10:12:07.292487+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14156 ' entity='mgr.a' 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: audit 2026-03-07T10:12:07.297204+0000 mon.b (mon.1) 12 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: audit 2026-03-07T10:12:07.297204+0000 mon.b (mon.1) 12 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: audit 2026-03-07T10:12:07.298004+0000 mon.b (mon.1) 13 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: audit 2026-03-07T10:12:07.298004+0000 mon.b (mon.1) 13 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: audit 2026-03-07T10:12:07.298717+0000 mon.b (mon.1) 14 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:12:07.582 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: audit 2026-03-07T10:12:07.298717+0000 mon.b (mon.1) 14 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: audit 2026-03-07T10:12:06.272624+0000 mon.b (mon.1) 10 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: audit 2026-03-07T10:12:06.272624+0000 mon.b (mon.1) 10 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: audit 2026-03-07T10:12:06.272972+0000 mon.b (mon.1) 11 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: audit 2026-03-07T10:12:06.272972+0000 mon.b (mon.1) 11 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:06.290579+0000 mgr.a (mgr.14156) 160 : cluster [DBG] pgmap v109: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:06.290579+0000 mgr.a (mgr.14156) 160 : cluster [DBG] pgmap v109: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:07.266454+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:07.266454+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:07.269253+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:07.269253+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:07.273462+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:07.273462+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:07.273471+0000 mon.a (mon.0) 4 : cluster [DBG] fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:07.273471+0000 mon.a (mon.0) 4 : cluster [DBG] fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:07.273476+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-07T10:09:14.324495+0000 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:07.273476+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-07T10:09:14.324495+0000 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:07.273480+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-07T10:07:51.937005+0000 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:07.273480+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-07T10:07:51.937005+0000 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:07.273484+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 19 (squid) 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:07.273484+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 19 (squid) 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:07.273489+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:07.273489+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:07.273493+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:07.273493+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:07.273498+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon.b 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:07.273498+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon.b 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:07.273870+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:07.273870+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:07.273893+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:07.273893+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:07.274579+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e14: a(active, since 3m), standbys: b 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:07.274579+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e14: a(active, since 3m), standbys: b 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:07.274687+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: cluster 2026-03-07T10:12:07.274687+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: audit 2026-03-07T10:12:07.283586+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14156 ' entity='' 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: audit 2026-03-07T10:12:07.283586+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14156 ' entity='' 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: audit 2026-03-07T10:12:07.292487+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14156 ' entity='mgr.a' 2026-03-07T10:12:07.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: audit 2026-03-07T10:12:07.292487+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14156 ' entity='mgr.a' 2026-03-07T10:12:07.713 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: audit 2026-03-07T10:12:07.297204+0000 mon.b (mon.1) 12 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T10:12:07.713 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: audit 2026-03-07T10:12:07.297204+0000 mon.b (mon.1) 12 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T10:12:07.713 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: audit 2026-03-07T10:12:07.298004+0000 mon.b (mon.1) 13 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-07T10:12:07.713 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: audit 2026-03-07T10:12:07.298004+0000 mon.b (mon.1) 13 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-07T10:12:07.713 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: audit 2026-03-07T10:12:07.298717+0000 mon.b (mon.1) 14 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:12:07.713 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:07 vm02 bash[34286]: audit 2026-03-07T10:12:07.298717+0000 mon.b (mon.1) 14 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:12:07.832 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:12:07.832 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 systemd[1]: Stopping Ceph mon.b for 6c715b7a-1a0d-11f1-b180-89615ccd948e... 2026-03-07T10:12:07.832 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:12:07 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:12:07.832 INFO:journalctl@ceph.osd.1.vm10.stdout:Mar 07 10:12:07 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:12:08.097 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: debug 2026-03-07T10:12:07.864+0000 7f7c30a3d640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.b -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-07T10:12:08.097 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:07 vm10 bash[19283]: debug 2026-03-07T10:12:07.864+0000 7f7c30a3d640 -1 mon.b@1(peon) e2 *** Got Signal Terminated *** 2026-03-07T10:12:08.097 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27619]: ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e-mon-b 2026-03-07T10:12:08.097 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 systemd[1]: ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@mon.b.service: Deactivated successfully. 2026-03-07T10:12:08.097 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 systemd[1]: Stopped Ceph mon.b for 6c715b7a-1a0d-11f1-b180-89615ccd948e. 2026-03-07T10:12:08.360 INFO:journalctl@ceph.osd.1.vm10.stdout:Mar 07 10:12:08 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:12:08.360 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:12:08.360 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 systemd[1]: Started Ceph mon.b for 6c715b7a-1a0d-11f1-b180-89615ccd948e. 2026-03-07T10:12:08.360 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:12:08 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.384+0000 7f57f8294d80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.384+0000 7f57f8294d80 0 ceph version 19.2.3-39-g340d3c24fc6 (340d3c24fc6ae7529322dc7ccee6c6cb2589da0a) squid (stable), process ceph-mon, pid 8 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.384+0000 7f57f8294d80 0 pidfile_write: ignore empty --pid-file 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 0 load: jerasure load: lrc 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: RocksDB version: 7.9.2 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Git sha 0 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Compile date 2026-03-06 13:52:12 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: DB SUMMARY 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: DB Session ID: 0T7G8VOSCF2YHBDMF912 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: CURRENT file: CURRENT 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: MANIFEST file: MANIFEST-000010 size: 179 Bytes 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-b/store.db dir, Total Num: 1, files: 000008.sst 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-b/store.db: 000009.log size: 4874096 ; 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.error_if_exists: 0 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.create_if_missing: 0 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.paranoid_checks: 1 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.env: 0x55fcb80ccca0 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.info_log: 0x55fce7a2b700 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.statistics: (nil) 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.use_fsync: 0 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.max_log_file_size: 0 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.allow_fallocate: 1 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.use_direct_reads: 0 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-07T10:12:08.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.db_log_dir: 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.wal_dir: 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.write_buffer_manager: 0x55fce7a2f900 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.unordered_write: 0 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.row_cache: None 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.wal_filter: None 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.two_write_queues: 0 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.wal_compression: 0 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.atomic_flush: 0 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.log_readahead_size: 0 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-07T10:12:08.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.max_background_jobs: 2 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.max_background_compactions: -1 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.max_subcompactions: 1 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.max_open_files: -1 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Options.max_background_flushes: -1 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Compression algorithms supported: 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: kZSTD supported: 0 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: kXpressCompression supported: 0 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: kBZip2Compression supported: 0 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: kLZ4Compression supported: 1 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: kZlibCompression supported: 1 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: kSnappyCompression supported: 1 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.388+0000 7f57f8294d80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000010 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.merge_operator: 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.compaction_filter: None 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55fce79e1f40) 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: cache_index_and_filter_blocks: 1 2026-03-07T10:12:08.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: pin_top_level_index_and_filter: 1 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: index_type: 0 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: data_block_index_type: 0 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: index_shortening: 1 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: data_block_hash_table_util_ratio: 0.750000 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: checksum: 4 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: no_block_cache: 0 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: block_cache: 0x55fce7a511f0 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: block_cache_name: BinnedLRUCache 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: block_cache_options: 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: capacity : 536870912 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: num_shard_bits : 4 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: strict_capacity_limit : 0 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: high_pri_pool_ratio: 0.000 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: block_cache_compressed: (nil) 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: persistent_cache: (nil) 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: block_size: 4096 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: block_size_deviation: 10 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: block_restart_interval: 16 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: index_block_restart_interval: 1 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: metadata_block_size: 4096 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: partition_filters: 0 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: use_delta_encoding: 1 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: filter_policy: bloomfilter 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: whole_key_filtering: 1 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: verify_compression: 0 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: read_amp_bytes_per_bit: 0 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: format_version: 5 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: enable_index_compression: 1 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: block_align: 0 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: max_auto_readahead_size: 262144 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: prepopulate_block_cache: 0 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: initial_auto_readahead_size: 8192 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: num_file_reads_for_auto_readahead: 2 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.compression: NoCompression 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-07T10:12:08.702 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.num_levels: 7 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-07T10:12:08.703 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.inplace_update_support: 0 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.bloom_locality: 0 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.max_successive_merges: 0 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.ttl: 2592000 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.enable_blob_files: false 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.min_blob_size: 0 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.392+0000 7f57f8294d80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.396+0000 7f57f8294d80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000010 succeeded,manifest_file_number is 10, next_file_number is 12, last_sequence is 5, log_number is 5,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 5 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.396+0000 7f57f8294d80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 5 2026-03-07T10:12:08.704 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.396+0000 7f57f8294d80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 25b6bf60-5008-45b9-bf5d-ffc62770f950 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.396+0000 7f57f8294d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1772878328397751, "job": 1, "event": "recovery_started", "wal_files": [9]} 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.396+0000 7f57f8294d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #9 mode 2 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.408+0000 7f57f8294d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1772878328412132, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 13, "file_size": 3551171, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 6, "largest_seqno": 3469, "table_properties": {"data_size": 3539057, "index_size": 6977, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 4101, "raw_key_size": 41090, "raw_average_key_size": 25, "raw_value_size": 3506749, "raw_average_value_size": 2163, "num_data_blocks": 329, "num_entries": 1621, "num_filter_entries": 1621, "num_deletions": 4, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1772878328, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "25b6bf60-5008-45b9-bf5d-ffc62770f950", "db_session_id": "0T7G8VOSCF2YHBDMF912", "orig_file_number": 13, "seqno_to_time_mapping": "N/A"}} 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.408+0000 7f57f8294d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1772878328412602, "job": 1, "event": "recovery_finished"} 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.412+0000 7f57f8294d80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 15 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.412+0000 7f57f8294d80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-b/store.db/000009.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.412+0000 7f57f8294d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55fce7a52e00 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.412+0000 7f57f8294d80 4 rocksdb: DB pointer 0x55fce7b62000 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.416+0000 7f57ee05e640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.416+0000 7f57ee05e640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: ** DB Stats ** 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: ** Compaction Stats [default] ** 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: L0 2/0 3.39 MB 0.5 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 342.7 0.01 0.00 1 0.010 0 0 0.0 0.0 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: Sum 2/0 3.39 MB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 342.7 0.01 0.00 1 0.010 0 0 0.0 0.0 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 342.7 0.01 0.00 1 0.010 0 0 0.0 0.0 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: ** Compaction Stats [default] ** 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 342.7 0.01 0.00 1 0.010 0 0 0.0 0.0 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: Flush(GB): cumulative 0.003, interval 0.003 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: AddFile(Total Files): cumulative 0, interval 0 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: AddFile(L0 Files): cumulative 0, interval 0 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: AddFile(Keys): cumulative 0, interval 0 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: Cumulative compaction: 0.00 GB write, 149.38 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: Interval compaction: 0.00 GB write, 149.38 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: Block cache BinnedLRUCache@0x55fce7a511f0#8 capacity: 512.00 MB usage: 12.83 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 1.6e-05 secs_since: 0 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: Block cache entry stats(count,size,portion): FilterBlock(2,4.64 KB,0.000885129%) IndexBlock(2,8.19 KB,0.00156164%) Misc(1,0.00 KB,0%) 2026-03-07T10:12:08.705 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: ** File Read Latency Histogram By Level [default] ** 2026-03-07T10:12:08.706 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.416+0000 7f57f8294d80 0 starting mon.b rank 1 at public addrs [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] at bind addrs [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon_data /var/lib/ceph/mon/ceph-b fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:12:08.706 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.416+0000 7f57f8294d80 1 mon.b@-1(???) e2 preinit fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:12:08.706 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.416+0000 7f57f8294d80 0 mon.b@-1(???).mds e1 new map 2026-03-07T10:12:08.706 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.416+0000 7f57f8294d80 0 mon.b@-1(???).mds e1 print_map 2026-03-07T10:12:08.706 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: e1 2026-03-07T10:12:08.706 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: btime 2026-03-07T10:07:53:157919+0000 2026-03-07T10:12:08.706 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: enable_multiple, ever_enabled_multiple: 1,1 2026-03-07T10:12:08.706 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-07T10:12:08.706 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: legacy client fscid: -1 2026-03-07T10:12:08.706 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: 2026-03-07T10:12:08.706 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: No filesystems configured 2026-03-07T10:12:08.706 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.416+0000 7f57f8294d80 0 mon.b@-1(???).osd e14 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-07T10:12:08.706 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.416+0000 7f57f8294d80 0 mon.b@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-07T10:12:08.706 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.416+0000 7f57f8294d80 0 mon.b@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-07T10:12:08.706 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.416+0000 7f57f8294d80 0 mon.b@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-07T10:12:08.706 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:08 vm10 bash[27733]: debug 2026-03-07T10:12:08.416+0000 7f57f8294d80 1 mon.b@-1(???).paxosservice(auth 1..8) refresh upgraded, format 0 -> 3 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.290844+0000 mgr.a (mgr.14156) 163 : cluster [DBG] pgmap v110: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.290844+0000 mgr.a (mgr.14156) 163 : cluster [DBG] pgmap v110: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.523874+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.523874+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.525269+0000 mon.a (mon.0) 19 : cluster [INF] mon.a calling monitor election 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.525269+0000 mon.a (mon.0) 19 : cluster [INF] mon.a calling monitor election 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.527974+0000 mon.a (mon.0) 20 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.527974+0000 mon.a (mon.0) 20 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.532195+0000 mon.a (mon.0) 21 : cluster [DBG] monmap epoch 2 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.532195+0000 mon.a (mon.0) 21 : cluster [DBG] monmap epoch 2 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.532309+0000 mon.a (mon.0) 22 : cluster [DBG] fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.532309+0000 mon.a (mon.0) 22 : cluster [DBG] fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.532319+0000 mon.a (mon.0) 23 : cluster [DBG] last_changed 2026-03-07T10:09:14.324495+0000 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.532319+0000 mon.a (mon.0) 23 : cluster [DBG] last_changed 2026-03-07T10:09:14.324495+0000 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.532329+0000 mon.a (mon.0) 24 : cluster [DBG] created 2026-03-07T10:07:51.937005+0000 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.532329+0000 mon.a (mon.0) 24 : cluster [DBG] created 2026-03-07T10:07:51.937005+0000 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.532339+0000 mon.a (mon.0) 25 : cluster [DBG] min_mon_release 19 (squid) 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.532339+0000 mon.a (mon.0) 25 : cluster [DBG] min_mon_release 19 (squid) 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.532348+0000 mon.a (mon.0) 26 : cluster [DBG] election_strategy: 1 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.532348+0000 mon.a (mon.0) 26 : cluster [DBG] election_strategy: 1 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.532358+0000 mon.a (mon.0) 27 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.532358+0000 mon.a (mon.0) 27 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.532368+0000 mon.a (mon.0) 28 : cluster [DBG] 1: [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon.b 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.532368+0000 mon.a (mon.0) 28 : cluster [DBG] 1: [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon.b 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.532654+0000 mon.a (mon.0) 29 : cluster [DBG] fsmap 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.532654+0000 mon.a (mon.0) 29 : cluster [DBG] fsmap 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.532674+0000 mon.a (mon.0) 30 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.532674+0000 mon.a (mon.0) 30 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.532851+0000 mon.a (mon.0) 31 : cluster [DBG] mgrmap e14: a(active, since 3m), standbys: b 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.532851+0000 mon.a (mon.0) 31 : cluster [DBG] mgrmap e14: a(active, since 3m), standbys: b 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.532926+0000 mon.a (mon.0) 32 : cluster [INF] overall HEALTH_OK 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: cluster 2026-03-07T10:12:08.532926+0000 mon.a (mon.0) 32 : cluster [INF] overall HEALTH_OK 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: audit 2026-03-07T10:12:08.536483+0000 mon.a (mon.0) 33 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: audit 2026-03-07T10:12:08.536483+0000 mon.a (mon.0) 33 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: audit 2026-03-07T10:12:08.539663+0000 mon.a (mon.0) 34 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:09.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: audit 2026-03-07T10:12:08.539663+0000 mon.a (mon.0) 34 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:09.947 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: audit 2026-03-07T10:12:08.540505+0000 mon.a (mon.0) 35 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:12:09.947 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:09 vm10 bash[27733]: audit 2026-03-07T10:12:08.540505+0000 mon.a (mon.0) 35 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.290844+0000 mgr.a (mgr.14156) 163 : cluster [DBG] pgmap v110: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.290844+0000 mgr.a (mgr.14156) 163 : cluster [DBG] pgmap v110: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.523874+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.523874+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.525269+0000 mon.a (mon.0) 19 : cluster [INF] mon.a calling monitor election 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.525269+0000 mon.a (mon.0) 19 : cluster [INF] mon.a calling monitor election 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.527974+0000 mon.a (mon.0) 20 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.527974+0000 mon.a (mon.0) 20 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.532195+0000 mon.a (mon.0) 21 : cluster [DBG] monmap epoch 2 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.532195+0000 mon.a (mon.0) 21 : cluster [DBG] monmap epoch 2 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.532309+0000 mon.a (mon.0) 22 : cluster [DBG] fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.532309+0000 mon.a (mon.0) 22 : cluster [DBG] fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.532319+0000 mon.a (mon.0) 23 : cluster [DBG] last_changed 2026-03-07T10:09:14.324495+0000 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.532319+0000 mon.a (mon.0) 23 : cluster [DBG] last_changed 2026-03-07T10:09:14.324495+0000 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.532329+0000 mon.a (mon.0) 24 : cluster [DBG] created 2026-03-07T10:07:51.937005+0000 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.532329+0000 mon.a (mon.0) 24 : cluster [DBG] created 2026-03-07T10:07:51.937005+0000 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.532339+0000 mon.a (mon.0) 25 : cluster [DBG] min_mon_release 19 (squid) 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.532339+0000 mon.a (mon.0) 25 : cluster [DBG] min_mon_release 19 (squid) 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.532348+0000 mon.a (mon.0) 26 : cluster [DBG] election_strategy: 1 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.532348+0000 mon.a (mon.0) 26 : cluster [DBG] election_strategy: 1 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.532358+0000 mon.a (mon.0) 27 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.532358+0000 mon.a (mon.0) 27 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.532368+0000 mon.a (mon.0) 28 : cluster [DBG] 1: [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon.b 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.532368+0000 mon.a (mon.0) 28 : cluster [DBG] 1: [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon.b 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.532654+0000 mon.a (mon.0) 29 : cluster [DBG] fsmap 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.532654+0000 mon.a (mon.0) 29 : cluster [DBG] fsmap 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.532674+0000 mon.a (mon.0) 30 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.532674+0000 mon.a (mon.0) 30 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.532851+0000 mon.a (mon.0) 31 : cluster [DBG] mgrmap e14: a(active, since 3m), standbys: b 2026-03-07T10:12:09.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.532851+0000 mon.a (mon.0) 31 : cluster [DBG] mgrmap e14: a(active, since 3m), standbys: b 2026-03-07T10:12:09.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.532926+0000 mon.a (mon.0) 32 : cluster [INF] overall HEALTH_OK 2026-03-07T10:12:09.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: cluster 2026-03-07T10:12:08.532926+0000 mon.a (mon.0) 32 : cluster [INF] overall HEALTH_OK 2026-03-07T10:12:09.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: audit 2026-03-07T10:12:08.536483+0000 mon.a (mon.0) 33 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:09.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: audit 2026-03-07T10:12:08.536483+0000 mon.a (mon.0) 33 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:09.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: audit 2026-03-07T10:12:08.539663+0000 mon.a (mon.0) 34 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:09.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: audit 2026-03-07T10:12:08.539663+0000 mon.a (mon.0) 34 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:09.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: audit 2026-03-07T10:12:08.540505+0000 mon.a (mon.0) 35 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:12:09.963 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:09 vm02 bash[34286]: audit 2026-03-07T10:12:08.540505+0000 mon.a (mon.0) 35 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:12:10.683 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config 2026-03-07T10:12:11.020 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-07T10:12:11.020 INFO:teuthology.orchestra.run.vm02.stdout:[{"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-07T10:12:00.517521Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-07T10:11:40.613675Z", "last_refresh": "2026-03-07T10:12:05.485847Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-07T10:11:42.445590Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-07T10:11:40.609750Z", "last_refresh": "2026-03-07T10:12:05.485681Z", "running": 2, "size": 2}}, {"events": ["2026-03-07T10:09:30.906338Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm02=a", "vm10=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-07T10:09:30.085053Z", "last_refresh": "2026-03-07T10:12:05.485817Z", "running": 2, "size": 2}}, {"events": ["2026-03-07T10:11:41.004615Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-07T10:11:40.604541Z", "running": 0, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "8bccc98d839aa18345ec1336292d0452ca331737e49f12524f635044dcabcfe1", "container_image_name": "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:ffa52c72fad7bdd2657408de9cf8d87fc2c72f716d1a00277ba13f7c12b404e0", "last_refresh": "2026-03-07T10:12:05.485786Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-07T10:12:11.092 INFO:tasks.cephadm:container.bar has 2/2 2026-03-07T10:12:11.093 INFO:teuthology.run_tasks:Running task exec... 2026-03-07T10:12:11.095 INFO:teuthology.task.exec:Executing custom commands... 2026-03-07T10:12:11.096 INFO:teuthology.task.exec:Running commands on role host.a host ubuntu@vm02.local 2026-03-07T10:12:11.096 DEBUG:teuthology.orchestra.run.vm02:> sudo TESTDIR=/home/ubuntu/cephtest bash -c 'set -ex 2026-03-07T10:12:11.096 DEBUG:teuthology.orchestra.run.vm02:> FSID=$(/home/ubuntu/cephtest/cephadm shell -- ceph fsid) 2026-03-07T10:12:11.096 DEBUG:teuthology.orchestra.run.vm02:> sleep 60 2026-03-07T10:12:11.096 DEBUG:teuthology.orchestra.run.vm02:> # check extra container and entrypoint args written to mon unit run file 2026-03-07T10:12:11.096 DEBUG:teuthology.orchestra.run.vm02:> grep "\-\-cpus=2" /var/lib/ceph/$FSID/mon.*/unit.run 2026-03-07T10:12:11.096 DEBUG:teuthology.orchestra.run.vm02:> grep "\-\-debug_ms 10" /var/lib/ceph/$FSID/mon.*/unit.run 2026-03-07T10:12:11.096 DEBUG:teuthology.orchestra.run.vm02:> # check that custom container properly wrote content to file. 2026-03-07T10:12:11.096 DEBUG:teuthology.orchestra.run.vm02:> # This requires the custom config, extra container args, and 2026-03-07T10:12:11.096 DEBUG:teuthology.orchestra.run.vm02:> # entrypoint args to all be working in order for this to have 2026-03-07T10:12:11.096 DEBUG:teuthology.orchestra.run.vm02:> # been written. The container entrypoint was set up with custom_configs, 2026-03-07T10:12:11.096 DEBUG:teuthology.orchestra.run.vm02:> # the content and where to write to with the entrypoint args, and the mounting 2026-03-07T10:12:11.096 DEBUG:teuthology.orchestra.run.vm02:> # of the /etc/cephadm_testing dir with extra container args 2026-03-07T10:12:11.096 DEBUG:teuthology.orchestra.run.vm02:> grep "testing_custom_containers" /etc/cephadm_testing/testing.txt 2026-03-07T10:12:11.096 DEBUG:teuthology.orchestra.run.vm02:> # Verify that container bar'"'"'s init containers and primary container 2026-03-07T10:12:11.096 DEBUG:teuthology.orchestra.run.vm02:> # ran successfully 2026-03-07T10:12:11.096 DEBUG:teuthology.orchestra.run.vm02:> dir=$(find /var/lib/ceph/$FSID -maxdepth 1 -type d -name '"'"'container.bar.*'"'"') 2026-03-07T10:12:11.096 DEBUG:teuthology.orchestra.run.vm02:> test -n "$dir" 2026-03-07T10:12:11.096 DEBUG:teuthology.orchestra.run.vm02:> grep ok ${dir}/data/primary.txt 2026-03-07T10:12:11.096 DEBUG:teuthology.orchestra.run.vm02:> grep from=centos ${dir}/data/from.txt 2026-03-07T10:12:11.096 DEBUG:teuthology.orchestra.run.vm02:> test -s ${dir}/data/presized.dat 2026-03-07T10:12:11.096 DEBUG:teuthology.orchestra.run.vm02:> ' 2026-03-07T10:12:11.104 INFO:teuthology.orchestra.run.vm02.stderr:++ /home/ubuntu/cephtest/cephadm shell -- ceph fsid 2026-03-07T10:12:11.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:11 vm10 bash[27733]: cluster 2026-03-07T10:12:10.291084+0000 mgr.a (mgr.14156) 164 : cluster [DBG] pgmap v111: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:11.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:11 vm10 bash[27733]: cluster 2026-03-07T10:12:10.291084+0000 mgr.a (mgr.14156) 164 : cluster [DBG] pgmap v111: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:11.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:11 vm02 bash[34286]: cluster 2026-03-07T10:12:10.291084+0000 mgr.a (mgr.14156) 164 : cluster [DBG] pgmap v111: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:11.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:11 vm02 bash[34286]: cluster 2026-03-07T10:12:10.291084+0000 mgr.a (mgr.14156) 164 : cluster [DBG] pgmap v111: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:12.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:12 vm10 bash[27733]: audit 2026-03-07T10:12:11.017077+0000 mgr.a (mgr.14156) 165 : audit [DBG] from='client.34100 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T10:12:12.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:12 vm10 bash[27733]: audit 2026-03-07T10:12:11.017077+0000 mgr.a (mgr.14156) 165 : audit [DBG] from='client.34100 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T10:12:12.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:12 vm02 bash[34286]: audit 2026-03-07T10:12:11.017077+0000 mgr.a (mgr.14156) 165 : audit [DBG] from='client.34100 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T10:12:12.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:12 vm02 bash[34286]: audit 2026-03-07T10:12:11.017077+0000 mgr.a (mgr.14156) 165 : audit [DBG] from='client.34100 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T10:12:14.132 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:13 vm02 bash[34286]: cluster 2026-03-07T10:12:12.291309+0000 mgr.a (mgr.14156) 166 : cluster [DBG] pgmap v112: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:14.132 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:13 vm02 bash[34286]: cluster 2026-03-07T10:12:12.291309+0000 mgr.a (mgr.14156) 166 : cluster [DBG] pgmap v112: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:14.132 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:13 vm02 bash[34286]: audit 2026-03-07T10:12:12.891389+0000 mon.a (mon.0) 36 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:14.132 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:13 vm02 bash[34286]: audit 2026-03-07T10:12:12.891389+0000 mon.a (mon.0) 36 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:14.132 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:13 vm02 bash[34286]: audit 2026-03-07T10:12:12.896208+0000 mon.a (mon.0) 37 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:14.132 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:13 vm02 bash[34286]: audit 2026-03-07T10:12:12.896208+0000 mon.a (mon.0) 37 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:14.132 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:13 vm02 bash[34286]: audit 2026-03-07T10:12:13.575165+0000 mon.a (mon.0) 38 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:14.132 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:13 vm02 bash[34286]: audit 2026-03-07T10:12:13.575165+0000 mon.a (mon.0) 38 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:14.132 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:13 vm02 bash[34286]: audit 2026-03-07T10:12:13.579173+0000 mon.a (mon.0) 39 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:14.132 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:13 vm02 bash[34286]: audit 2026-03-07T10:12:13.579173+0000 mon.a (mon.0) 39 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:14.132 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:13 vm02 bash[34286]: audit 2026-03-07T10:12:13.580065+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:12:14.132 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:13 vm02 bash[34286]: audit 2026-03-07T10:12:13.580065+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:12:14.132 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:13 vm02 bash[34286]: audit 2026-03-07T10:12:13.580481+0000 mon.a (mon.0) 41 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:12:14.132 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:13 vm02 bash[34286]: audit 2026-03-07T10:12:13.580481+0000 mon.a (mon.0) 41 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:12:14.132 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:13 vm02 bash[34286]: audit 2026-03-07T10:12:13.583544+0000 mon.a (mon.0) 42 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:14.132 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:13 vm02 bash[34286]: audit 2026-03-07T10:12:13.583544+0000 mon.a (mon.0) 42 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:14.132 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:13 vm02 bash[34286]: audit 2026-03-07T10:12:13.593053+0000 mon.a (mon.0) 43 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T10:12:14.132 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:13 vm02 bash[34286]: audit 2026-03-07T10:12:13.593053+0000 mon.a (mon.0) 43 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T10:12:14.132 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:13 vm02 bash[34286]: audit 2026-03-07T10:12:13.593439+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-07T10:12:14.132 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:13 vm02 bash[34286]: audit 2026-03-07T10:12:13.593439+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-07T10:12:14.132 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:13 vm02 bash[34286]: audit 2026-03-07T10:12:13.593756+0000 mon.a (mon.0) 45 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:12:14.132 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:13 vm02 bash[34286]: audit 2026-03-07T10:12:13.593756+0000 mon.a (mon.0) 45 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:12:14.132 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:12:14.132 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 systemd[1]: Stopping Ceph mon.a for 6c715b7a-1a0d-11f1-b180-89615ccd948e... 2026-03-07T10:12:14.132 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:12:14 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:12:14.133 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 07 10:12:14 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:12:14.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:13 vm10 bash[27733]: cluster 2026-03-07T10:12:12.291309+0000 mgr.a (mgr.14156) 166 : cluster [DBG] pgmap v112: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:14.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:13 vm10 bash[27733]: cluster 2026-03-07T10:12:12.291309+0000 mgr.a (mgr.14156) 166 : cluster [DBG] pgmap v112: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:14.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:13 vm10 bash[27733]: audit 2026-03-07T10:12:12.891389+0000 mon.a (mon.0) 36 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:14.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:13 vm10 bash[27733]: audit 2026-03-07T10:12:12.891389+0000 mon.a (mon.0) 36 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:14.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:13 vm10 bash[27733]: audit 2026-03-07T10:12:12.896208+0000 mon.a (mon.0) 37 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:14.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:13 vm10 bash[27733]: audit 2026-03-07T10:12:12.896208+0000 mon.a (mon.0) 37 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:14.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:13 vm10 bash[27733]: audit 2026-03-07T10:12:13.575165+0000 mon.a (mon.0) 38 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:14.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:13 vm10 bash[27733]: audit 2026-03-07T10:12:13.575165+0000 mon.a (mon.0) 38 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:14.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:13 vm10 bash[27733]: audit 2026-03-07T10:12:13.579173+0000 mon.a (mon.0) 39 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:14.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:13 vm10 bash[27733]: audit 2026-03-07T10:12:13.579173+0000 mon.a (mon.0) 39 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:14.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:13 vm10 bash[27733]: audit 2026-03-07T10:12:13.580065+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:12:14.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:13 vm10 bash[27733]: audit 2026-03-07T10:12:13.580065+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:12:14.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:13 vm10 bash[27733]: audit 2026-03-07T10:12:13.580481+0000 mon.a (mon.0) 41 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:12:14.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:13 vm10 bash[27733]: audit 2026-03-07T10:12:13.580481+0000 mon.a (mon.0) 41 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:12:14.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:13 vm10 bash[27733]: audit 2026-03-07T10:12:13.583544+0000 mon.a (mon.0) 42 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:14.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:13 vm10 bash[27733]: audit 2026-03-07T10:12:13.583544+0000 mon.a (mon.0) 42 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:14.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:13 vm10 bash[27733]: audit 2026-03-07T10:12:13.593053+0000 mon.a (mon.0) 43 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T10:12:14.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:13 vm10 bash[27733]: audit 2026-03-07T10:12:13.593053+0000 mon.a (mon.0) 43 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T10:12:14.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:13 vm10 bash[27733]: audit 2026-03-07T10:12:13.593439+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-07T10:12:14.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:13 vm10 bash[27733]: audit 2026-03-07T10:12:13.593439+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-07T10:12:14.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:13 vm10 bash[27733]: audit 2026-03-07T10:12:13.593756+0000 mon.a (mon.0) 45 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:12:14.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:13 vm10 bash[27733]: audit 2026-03-07T10:12:13.593756+0000 mon.a (mon.0) 45 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:12:14.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34286]: debug 2026-03-07T10:12:14.157+0000 7f9f9c26e640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-07T10:12:14.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34286]: debug 2026-03-07T10:12:14.157+0000 7f9f9c26e640 -1 mon.a@0(leader) e2 *** Got Signal Terminated *** 2026-03-07T10:12:14.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34753]: ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e-mon-a 2026-03-07T10:12:14.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 systemd[1]: ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@mon.a.service: Deactivated successfully. 2026-03-07T10:12:14.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 systemd[1]: Stopped Ceph mon.a for 6c715b7a-1a0d-11f1-b180-89615ccd948e. 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 07 10:12:14 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:12:14.836 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 07 10:12:14 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 systemd[1]: Started Ceph mon.a for 6c715b7a-1a0d-11f1-b180-89615ccd948e. 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.645+0000 7f955c5a6d80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.645+0000 7f955c5a6d80 0 ceph version 19.2.3-39-g340d3c24fc6 (340d3c24fc6ae7529322dc7ccee6c6cb2589da0a) squid (stable), process ceph-mon, pid 7 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.645+0000 7f955c5a6d80 0 pidfile_write: ignore empty --pid-file 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 0 load: jerasure load: lrc 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: RocksDB version: 7.9.2 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Git sha 0 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Compile date 2026-03-06 13:52:12 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: DB SUMMARY 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: DB Session ID: UCP719MAS1625SI3JH0Z 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: CURRENT file: CURRENT 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: MANIFEST file: MANIFEST-000020 size: 373 Bytes 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 3, files: 000008.sst 000013.sst 000018.sst 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 000019.log size: 250850 ; 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.error_if_exists: 0 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.create_if_missing: 0 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.paranoid_checks: 1 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.env: 0x5622afc74ca0 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.info_log: 0x5622d1755700 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.statistics: (nil) 2026-03-07T10:12:14.836 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.use_fsync: 0 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.max_log_file_size: 0 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.allow_fallocate: 1 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.use_direct_reads: 0 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.db_log_dir: 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.wal_dir: 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.write_buffer_manager: 0x5622d1759900 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.unordered_write: 0 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.row_cache: None 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.wal_filter: None 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-07T10:12:14.837 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.two_write_queues: 0 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.wal_compression: 0 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.atomic_flush: 0 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.log_readahead_size: 0 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.max_background_jobs: 2 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.max_background_compactions: -1 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.max_subcompactions: 1 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.max_open_files: -1 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.max_background_flushes: -1 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Compression algorithms supported: 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: kZSTD supported: 0 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: kXpressCompression supported: 0 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: kBZip2Compression supported: 0 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: kLZ4Compression supported: 1 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: kZlibCompression supported: 1 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: kSnappyCompression supported: 1 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000020 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.merge_operator: 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.compaction_filter: None 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5622d1754080) 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: cache_index_and_filter_blocks: 1 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: pin_top_level_index_and_filter: 1 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: index_type: 0 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: data_block_index_type: 0 2026-03-07T10:12:14.838 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: index_shortening: 1 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: data_block_hash_table_util_ratio: 0.750000 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: checksum: 4 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: no_block_cache: 0 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: block_cache: 0x5622d177b1f0 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: block_cache_name: BinnedLRUCache 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: block_cache_options: 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: capacity : 536870912 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: num_shard_bits : 4 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: strict_capacity_limit : 0 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: high_pri_pool_ratio: 0.000 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: block_cache_compressed: (nil) 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: persistent_cache: (nil) 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: block_size: 4096 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: block_size_deviation: 10 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: block_restart_interval: 16 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: index_block_restart_interval: 1 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: metadata_block_size: 4096 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: partition_filters: 0 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: use_delta_encoding: 1 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: filter_policy: bloomfilter 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: whole_key_filtering: 1 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: verify_compression: 0 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: read_amp_bytes_per_bit: 0 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: format_version: 5 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: enable_index_compression: 1 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: block_align: 0 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: max_auto_readahead_size: 262144 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: prepopulate_block_cache: 0 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: initial_auto_readahead_size: 8192 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: num_file_reads_for_auto_readahead: 2 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.compression: NoCompression 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.num_levels: 7 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-07T10:12:14.839 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.inplace_update_support: 0 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.bloom_locality: 0 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.max_successive_merges: 0 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.ttl: 2592000 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.enable_blob_files: false 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.min_blob_size: 0 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000020 succeeded,manifest_file_number is 20, next_file_number is 22, last_sequence is 3499, log_number is 15,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 15 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 15 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: ab4b5d8f-62d9-446c-92eb-b6ba735315c3 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1772878334655448, "job": 1, "event": "recovery_started", "wal_files": [19]} 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.649+0000 7f955c5a6d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #19 mode 2 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.653+0000 7f955c5a6d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1772878334657255, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 23, "file_size": 215532, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 3505, "largest_seqno": 3709, "table_properties": {"data_size": 213388, "index_size": 791, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 325, "raw_key_size": 2786, "raw_average_key_size": 24, "raw_value_size": 210783, "raw_average_value_size": 1865, "num_data_blocks": 34, "num_entries": 113, "num_filter_entries": 113, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1772878334, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "ab4b5d8f-62d9-446c-92eb-b6ba735315c3", "db_session_id": "UCP719MAS1625SI3JH0Z", "orig_file_number": 23, "seqno_to_time_mapping": "N/A"}} 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.653+0000 7f955c5a6d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1772878334657304, "job": 1, "event": "recovery_finished"} 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.653+0000 7f955c5a6d80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 25 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.657+0000 7f955c5a6d80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-a/store.db/000019.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.657+0000 7f955c5a6d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x5622d177ce00 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.657+0000 7f955c5a6d80 4 rocksdb: DB pointer 0x5622d188c000 2026-03-07T10:12:14.840 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.657+0000 7f955c5a6d80 10 CompressorRegistry(0x5622d1743418) _refresh_config ms_osd_compression_mode 0 ms_osd_compression_methods [1] ms_osd_compress_above_min_size 1024 ms_compress_secure 0 2026-03-07T10:12:14.841 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.657+0000 7f955937e640 2 Event(0x5622d27402c8 nevent=5000 time_id=1).set_owner center_id=2 owner=140279423690304 2026-03-07T10:12:14.841 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.657+0000 7f955937e640 10 stack operator() starting 2026-03-07T10:12:14.841 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.657+0000 7f9553b73640 2 Event(0x5622d2740088 nevent=5000 time_id=1).set_owner center_id=1 owner=140279331370560 2026-03-07T10:12:14.841 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.657+0000 7f9553b73640 10 stack operator() starting 2026-03-07T10:12:14.841 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.657+0000 7f9553372640 2 Event(0x5622d1775b88 nevent=5000 time_id=1).set_owner center_id=0 owner=140279322977856 2026-03-07T10:12:14.841 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.657+0000 7f9553372640 10 stack operator() starting 2026-03-07T10:12:14.841 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.657+0000 7f955c5a6d80 0 starting mon.a rank 0 at public addrs [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] at bind addrs [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon_data /var/lib/ceph/mon/ceph-a fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:12:14.841 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.657+0000 7f955c5a6d80 10 CompressorRegistry(0x5622d1743d18) _refresh_config ms_osd_compression_mode 0 ms_osd_compression_methods [1] ms_osd_compress_above_min_size 1024 ms_compress_secure 0 2026-03-07T10:12:14.841 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.657+0000 7f955c5a6d80 1 mon.a@-1(???) e2 preinit fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:12:14.841 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.657+0000 7f955c5a6d80 0 mon.a@-1(???).mds e1 new map 2026-03-07T10:12:14.841 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.657+0000 7f955c5a6d80 0 mon.a@-1(???).mds e1 print_map 2026-03-07T10:12:14.841 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: e1 2026-03-07T10:12:14.841 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: btime 2026-03-07T10:07:53:157919+0000 2026-03-07T10:12:14.841 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: enable_multiple, ever_enabled_multiple: 1,1 2026-03-07T10:12:14.841 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-07T10:12:14.841 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: legacy client fscid: -1 2026-03-07T10:12:14.841 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: 2026-03-07T10:12:14.841 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: No filesystems configured 2026-03-07T10:12:14.841 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.657+0000 7f955c5a6d80 0 mon.a@-1(???).osd e14 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-07T10:12:14.841 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.657+0000 7f955c5a6d80 0 mon.a@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-07T10:12:14.841 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.657+0000 7f955c5a6d80 0 mon.a@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-07T10:12:14.841 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.657+0000 7f955c5a6d80 0 mon.a@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-07T10:12:14.841 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.657+0000 7f955c5a6d80 1 mon.a@-1(???).paxosservice(auth 1..9) refresh upgraded, format 0 -> 3 2026-03-07T10:12:14.841 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.657+0000 7f955a380640 4 rocksdb: [db/compaction/compaction_job.cc:1995] [default] [JOB 3] Compacting 4@0 files to L6, score 1.00 2026-03-07T10:12:14.841 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.657+0000 7f955a380640 4 rocksdb: [db/compaction/compaction_job.cc:2001] [default]: Compaction start summary: Base version 2 Base level 0, inputs: [23(210KB) 18(3357KB) 13(71KB) 8(1900B)] 2026-03-07T10:12:14.841 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:14 vm02 bash[34866]: debug 2026-03-07T10:12:14.657+0000 7f955a380640 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1772878334663256, "job": 3, "event": "compaction_started", "compaction_reason": "LevelL0FilesNum", "files_L0": [23, 18, 13, 8], "score": 1, "input_data_size": 3728437, "oldest_snapshot_seqno": -1} 2026-03-07T10:12:16.394 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: audit 2026-03-07T10:12:14.173658+0000 mon.b (mon.1) 2 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-07T10:12:16.394 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: audit 2026-03-07T10:12:14.173658+0000 mon.b (mon.1) 2 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-07T10:12:16.394 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: audit 2026-03-07T10:12:14.173973+0000 mon.b (mon.1) 3 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:12:16.394 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: audit 2026-03-07T10:12:14.173973+0000 mon.b (mon.1) 3 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:12:16.394 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:14.291548+0000 mgr.a (mgr.14156) 169 : cluster [DBG] pgmap v113: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:16.394 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:14.291548+0000 mgr.a (mgr.14156) 169 : cluster [DBG] pgmap v113: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:16.394 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:16.071969+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-07T10:12:16.394 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:16.071969+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-07T10:12:16.394 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:16.075474+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-07T10:12:16.394 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:16.075474+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-07T10:12:16.394 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:16.079506+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-07T10:12:16.394 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:16.079506+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-07T10:12:16.394 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:16.079577+0000 mon.a (mon.0) 4 : cluster [DBG] fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:12:16.394 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:16.079577+0000 mon.a (mon.0) 4 : cluster [DBG] fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:12:16.394 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:16.079587+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-07T10:09:14.324495+0000 2026-03-07T10:12:16.394 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:16.079587+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-07T10:09:14.324495+0000 2026-03-07T10:12:16.394 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:16.079595+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-07T10:07:51.937005+0000 2026-03-07T10:12:16.394 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:16.079595+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-07T10:07:51.937005+0000 2026-03-07T10:12:16.394 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:16.079605+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 19 (squid) 2026-03-07T10:12:16.394 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:16.079605+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 19 (squid) 2026-03-07T10:12:16.394 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:16.079613+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-07T10:12:16.394 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:16.079613+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-07T10:12:16.394 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:16.079640+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-07T10:12:16.395 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:16.079640+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-07T10:12:16.395 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:16.079649+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon.b 2026-03-07T10:12:16.395 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:16.079649+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon.b 2026-03-07T10:12:16.395 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:16.079991+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-07T10:12:16.395 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:16.079991+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-07T10:12:16.395 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:16.080071+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-07T10:12:16.395 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:16.080071+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-07T10:12:16.395 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:16.080842+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e14: a(active, since 3m), standbys: b 2026-03-07T10:12:16.395 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:16.080842+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e14: a(active, since 3m), standbys: b 2026-03-07T10:12:16.395 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:16.081005+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-07T10:12:16.395 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: cluster 2026-03-07T10:12:16.081005+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-07T10:12:16.395 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: audit 2026-03-07T10:12:16.089016+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14156 ' entity='' 2026-03-07T10:12:16.395 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: audit 2026-03-07T10:12:16.089016+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14156 ' entity='' 2026-03-07T10:12:16.395 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: audit 2026-03-07T10:12:16.095731+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14156 ' entity='mgr.a' 2026-03-07T10:12:16.395 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: audit 2026-03-07T10:12:16.095731+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14156 ' entity='mgr.a' 2026-03-07T10:12:16.395 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: audit 2026-03-07T10:12:16.099155+0000 mon.b (mon.1) 4 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T10:12:16.395 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: audit 2026-03-07T10:12:16.099155+0000 mon.b (mon.1) 4 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T10:12:16.395 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: audit 2026-03-07T10:12:16.100001+0000 mon.b (mon.1) 5 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-07T10:12:16.395 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: audit 2026-03-07T10:12:16.100001+0000 mon.b (mon.1) 5 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-07T10:12:16.395 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: audit 2026-03-07T10:12:16.100887+0000 mon.b (mon.1) 6 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:12:16.395 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: audit 2026-03-07T10:12:16.100887+0000 mon.b (mon.1) 6 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: audit 2026-03-07T10:12:14.173658+0000 mon.b (mon.1) 2 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: audit 2026-03-07T10:12:14.173658+0000 mon.b (mon.1) 2 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: audit 2026-03-07T10:12:14.173973+0000 mon.b (mon.1) 3 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: audit 2026-03-07T10:12:14.173973+0000 mon.b (mon.1) 3 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:14.291548+0000 mgr.a (mgr.14156) 169 : cluster [DBG] pgmap v113: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:14.291548+0000 mgr.a (mgr.14156) 169 : cluster [DBG] pgmap v113: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:16.071969+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:16.071969+0000 mon.a (mon.0) 1 : cluster [INF] mon.a calling monitor election 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:16.075474+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:16.075474+0000 mon.a (mon.0) 2 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:16.079506+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:16.079506+0000 mon.a (mon.0) 3 : cluster [DBG] monmap epoch 2 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:16.079577+0000 mon.a (mon.0) 4 : cluster [DBG] fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:16.079577+0000 mon.a (mon.0) 4 : cluster [DBG] fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:16.079587+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-07T10:09:14.324495+0000 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:16.079587+0000 mon.a (mon.0) 5 : cluster [DBG] last_changed 2026-03-07T10:09:14.324495+0000 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:16.079595+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-07T10:07:51.937005+0000 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:16.079595+0000 mon.a (mon.0) 6 : cluster [DBG] created 2026-03-07T10:07:51.937005+0000 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:16.079605+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 19 (squid) 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:16.079605+0000 mon.a (mon.0) 7 : cluster [DBG] min_mon_release 19 (squid) 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:16.079613+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:16.079613+0000 mon.a (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:16.079640+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:16.079640+0000 mon.a (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:16.079649+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon.b 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:16.079649+0000 mon.a (mon.0) 10 : cluster [DBG] 1: [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon.b 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:16.079991+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:16.079991+0000 mon.a (mon.0) 11 : cluster [DBG] fsmap 2026-03-07T10:12:16.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:16.080071+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-07T10:12:16.463 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:16.080071+0000 mon.a (mon.0) 12 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-07T10:12:16.463 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:16.080842+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e14: a(active, since 3m), standbys: b 2026-03-07T10:12:16.463 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:16.080842+0000 mon.a (mon.0) 13 : cluster [DBG] mgrmap e14: a(active, since 3m), standbys: b 2026-03-07T10:12:16.463 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:16.081005+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-07T10:12:16.463 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: cluster 2026-03-07T10:12:16.081005+0000 mon.a (mon.0) 14 : cluster [INF] overall HEALTH_OK 2026-03-07T10:12:16.463 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: audit 2026-03-07T10:12:16.089016+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14156 ' entity='' 2026-03-07T10:12:16.463 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: audit 2026-03-07T10:12:16.089016+0000 mon.a (mon.0) 15 : audit [INF] from='mgr.14156 ' entity='' 2026-03-07T10:12:16.463 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: audit 2026-03-07T10:12:16.095731+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14156 ' entity='mgr.a' 2026-03-07T10:12:16.463 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: audit 2026-03-07T10:12:16.095731+0000 mon.a (mon.0) 16 : audit [INF] from='mgr.14156 ' entity='mgr.a' 2026-03-07T10:12:16.463 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: audit 2026-03-07T10:12:16.099155+0000 mon.b (mon.1) 4 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T10:12:16.463 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: audit 2026-03-07T10:12:16.099155+0000 mon.b (mon.1) 4 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T10:12:16.463 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: audit 2026-03-07T10:12:16.100001+0000 mon.b (mon.1) 5 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-07T10:12:16.463 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: audit 2026-03-07T10:12:16.100001+0000 mon.b (mon.1) 5 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-07T10:12:16.463 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: audit 2026-03-07T10:12:16.100887+0000 mon.b (mon.1) 6 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:12:16.463 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:16 vm02 bash[34866]: audit 2026-03-07T10:12:16.100887+0000 mon.b (mon.1) 6 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:12:16.645 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:12:16.645 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 systemd[1]: Stopping Ceph mon.b for 6c715b7a-1a0d-11f1-b180-89615ccd948e... 2026-03-07T10:12:16.645 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:12:16 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:12:16.645 INFO:journalctl@ceph.osd.1.vm10.stdout:Mar 07 10:12:16 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:12:16.703 INFO:teuthology.orchestra.run.vm02.stderr:Inferring fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:12:16.901 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: debug 2026-03-07T10:12:16.676+0000 7f57f686f640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.b -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-07T10:12:16.901 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[27733]: debug 2026-03-07T10:12:16.676+0000 7f57f686f640 -1 mon.b@1(peon) e2 *** Got Signal Terminated *** 2026-03-07T10:12:17.195 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:12:17 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:12:17.196 INFO:journalctl@ceph.osd.1.vm10.stdout:Mar 07 10:12:17 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:12:17.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 bash[28076]: ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e-mon-b 2026-03-07T10:12:17.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 systemd[1]: ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@mon.b.service: Deactivated successfully. 2026-03-07T10:12:17.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:16 vm10 systemd[1]: Stopped Ceph mon.b for 6c715b7a-1a0d-11f1-b180-89615ccd948e. 2026-03-07T10:12:17.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 systemd[1]: /etc/systemd/system/ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T10:12:17.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 systemd[1]: Started Ceph mon.b for 6c715b7a-1a0d-11f1-b180-89615ccd948e. 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.244+0000 7f9533f85d80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.244+0000 7f9533f85d80 0 ceph version 19.2.3-39-g340d3c24fc6 (340d3c24fc6ae7529322dc7ccee6c6cb2589da0a) squid (stable), process ceph-mon, pid 7 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.244+0000 7f9533f85d80 0 pidfile_write: ignore empty --pid-file 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.244+0000 7f9533f85d80 0 load: jerasure load: lrc 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: RocksDB version: 7.9.2 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Git sha 0 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Compile date 2026-03-06 13:52:12 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: DB SUMMARY 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: DB Session ID: 1QA5YEIUNG87HB92PFKO 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: CURRENT file: CURRENT 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: MANIFEST file: MANIFEST-000015 size: 282 Bytes 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-b/store.db dir, Total Num: 2, files: 000008.sst 000013.sst 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-b/store.db: 000014.log size: 366926 ; 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.error_if_exists: 0 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.create_if_missing: 0 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.paranoid_checks: 1 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.env: 0x561661a78ca0 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.info_log: 0x561674915700 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.statistics: (nil) 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.use_fsync: 0 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.max_log_file_size: 0 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-07T10:12:17.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.allow_fallocate: 1 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.use_direct_reads: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.db_log_dir: 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.wal_dir: 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.write_buffer_manager: 0x561674919900 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.unordered_write: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.row_cache: None 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.wal_filter: None 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.two_write_queues: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.wal_compression: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.atomic_flush: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.log_readahead_size: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.max_background_jobs: 2 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.max_background_compactions: -1 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.max_subcompactions: 1 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.max_open_files: -1 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.max_background_flushes: -1 2026-03-07T10:12:17.697 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Compression algorithms supported: 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: kZSTD supported: 0 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: kXpressCompression supported: 0 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: kBZip2Compression supported: 0 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: kLZ4Compression supported: 1 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: kZlibCompression supported: 1 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: kSnappyCompression supported: 1 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000015 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.merge_operator: 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.compaction_filter: None 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5616748cbf40) 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: cache_index_and_filter_blocks: 1 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: pin_top_level_index_and_filter: 1 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: index_type: 0 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: data_block_index_type: 0 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: index_shortening: 1 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: data_block_hash_table_util_ratio: 0.750000 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: checksum: 4 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: no_block_cache: 0 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: block_cache: 0x56167493b1f0 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: block_cache_name: BinnedLRUCache 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: block_cache_options: 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: capacity : 536870912 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: num_shard_bits : 4 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: strict_capacity_limit : 0 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: high_pri_pool_ratio: 0.000 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: block_cache_compressed: (nil) 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: persistent_cache: (nil) 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: block_size: 4096 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: block_size_deviation: 10 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: block_restart_interval: 16 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: index_block_restart_interval: 1 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: metadata_block_size: 4096 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: partition_filters: 0 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: use_delta_encoding: 1 2026-03-07T10:12:17.698 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: filter_policy: bloomfilter 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: whole_key_filtering: 1 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: verify_compression: 0 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: read_amp_bytes_per_bit: 0 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: format_version: 5 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: enable_index_compression: 1 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: block_align: 0 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: max_auto_readahead_size: 262144 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: prepopulate_block_cache: 0 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: initial_auto_readahead_size: 8192 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: num_file_reads_for_auto_readahead: 2 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.compression: NoCompression 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.num_levels: 7 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-07T10:12:17.699 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.inplace_update_support: 0 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.bloom_locality: 0 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.max_successive_merges: 0 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.ttl: 2592000 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.enable_blob_files: false 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.min_blob_size: 0 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000015 succeeded,manifest_file_number is 15, next_file_number is 17, last_sequence is 3469, log_number is 10,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 10 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 10 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 25b6bf60-5008-45b9-bf5d-ffc62770f950 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1772878337250877, "job": 1, "event": "recovery_started", "wal_files": [14]} 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.248+0000 7f9533f85d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #14 mode 2 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.256+0000 7f9533f85d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1772878337257306, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 18, "file_size": 217619, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 3481, "largest_seqno": 3696, "table_properties": {"data_size": 215507, "index_size": 759, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 325, "raw_key_size": 2836, "raw_average_key_size": 24, "raw_value_size": 212867, "raw_average_value_size": 1851, "num_data_blocks": 33, "num_entries": 115, "num_filter_entries": 115, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1772878337, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "25b6bf60-5008-45b9-bf5d-ffc62770f950", "db_session_id": "1QA5YEIUNG87HB92PFKO", "orig_file_number": 18, "seqno_to_time_mapping": "N/A"}} 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.256+0000 7f9533f85d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1772878337257365, "job": 1, "event": "recovery_finished"} 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.256+0000 7f9533f85d80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 20 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.256+0000 7f9533f85d80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-b/store.db/000014.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.256+0000 7f9533f85d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x56167493ce00 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.256+0000 7f9533f85d80 4 rocksdb: DB pointer 0x561674a4c000 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.256+0000 7f9533f85d80 10 CompressorRegistry(0x561674904618) _refresh_config ms_osd_compression_mode 0 ms_osd_compression_methods [1] ms_osd_compress_above_min_size 1024 ms_compress_secure 0 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.260+0000 7f952ad51640 2 Event(0x561674935b88 nevent=5000 time_id=1).set_owner center_id=0 owner=140278645462592 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.260+0000 7f952ad51640 10 stack operator() starting 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.260+0000 7f952b552640 2 Event(0x561675992088 nevent=5000 time_id=1).set_owner center_id=1 owner=140278653855296 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.260+0000 7f952b552640 10 stack operator() starting 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.260+0000 7f9529d4f640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.260+0000 7f9529d4f640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: ** DB Stats ** 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: ** Compaction Stats [default] ** 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: L0 3/0 3.60 MB 0.8 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 154.8 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: Sum 3/0 3.60 MB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 154.8 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 154.8 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: ** Compaction Stats [default] ** 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 154.8 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-07T10:12:17.700 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: Flush(GB): cumulative 0.000, interval 0.000 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: AddFile(Total Files): cumulative 0, interval 0 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: AddFile(L0 Files): cumulative 0, interval 0 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: AddFile(Keys): cumulative 0, interval 0 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: Cumulative compaction: 0.00 GB write, 17.88 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: Interval compaction: 0.00 GB write, 17.88 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: Block cache BinnedLRUCache@0x56167493b1f0#7 capacity: 512.00 MB usage: 36.91 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 6e-06 secs_since: 0 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: Block cache entry stats(count,size,portion): DataBlock(4,22.88 KB,0.00436306%) FilterBlock(3,5.02 KB,0.000956655%) IndexBlock(3,9.02 KB,0.00171959%) Misc(1,0.00 KB,0%) 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: ** File Read Latency Histogram By Level [default] ** 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.260+0000 7f9530d5d640 2 Event(0x5616759922c8 nevent=5000 time_id=1).set_owner center_id=2 owner=140278746175040 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.260+0000 7f9530d5d640 10 stack operator() starting 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.260+0000 7f9533f85d80 0 starting mon.b rank 1 at public addrs [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] at bind addrs [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon_data /var/lib/ceph/mon/ceph-b fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.260+0000 7f9533f85d80 10 CompressorRegistry(0x561674904f18) _refresh_config ms_osd_compression_mode 0 ms_osd_compression_methods [1] ms_osd_compress_above_min_size 1024 ms_compress_secure 0 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.260+0000 7f9533f85d80 1 mon.b@-1(???) e2 preinit fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.260+0000 7f9533f85d80 0 mon.b@-1(???).mds e1 new map 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.260+0000 7f9533f85d80 0 mon.b@-1(???).mds e1 print_map 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: e1 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: btime 2026-03-07T10:07:53:157919+0000 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: enable_multiple, ever_enabled_multiple: 1,1 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: legacy client fscid: -1 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: No filesystems configured 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.260+0000 7f9533f85d80 0 mon.b@-1(???).osd e14 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.260+0000 7f9533f85d80 0 mon.b@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.260+0000 7f9533f85d80 0 mon.b@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.260+0000 7f9533f85d80 0 mon.b@-1(???).osd e14 crush map has features 288514050185494528, adjusting msgr requires 2026-03-07T10:12:17.701 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:17 vm10 bash[28189]: debug 2026-03-07T10:12:17.260+0000 7f9533f85d80 1 mon.b@-1(???).paxosservice(auth 1..10) refresh upgraded, format 0 -> 3 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.352298+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.352298+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.352793+0000 mon.a (mon.0) 19 : cluster [INF] mon.a calling monitor election 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.352793+0000 mon.a (mon.0) 19 : cluster [INF] mon.a calling monitor election 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.354824+0000 mon.a (mon.0) 20 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.354824+0000 mon.a (mon.0) 20 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.358794+0000 mon.a (mon.0) 21 : cluster [DBG] monmap epoch 2 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.358794+0000 mon.a (mon.0) 21 : cluster [DBG] monmap epoch 2 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.358871+0000 mon.a (mon.0) 22 : cluster [DBG] fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.358871+0000 mon.a (mon.0) 22 : cluster [DBG] fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.358894+0000 mon.a (mon.0) 23 : cluster [DBG] last_changed 2026-03-07T10:09:14.324495+0000 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.358894+0000 mon.a (mon.0) 23 : cluster [DBG] last_changed 2026-03-07T10:09:14.324495+0000 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.358910+0000 mon.a (mon.0) 24 : cluster [DBG] created 2026-03-07T10:07:51.937005+0000 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.358910+0000 mon.a (mon.0) 24 : cluster [DBG] created 2026-03-07T10:07:51.937005+0000 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.358927+0000 mon.a (mon.0) 25 : cluster [DBG] min_mon_release 19 (squid) 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.358927+0000 mon.a (mon.0) 25 : cluster [DBG] min_mon_release 19 (squid) 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.358942+0000 mon.a (mon.0) 26 : cluster [DBG] election_strategy: 1 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.358942+0000 mon.a (mon.0) 26 : cluster [DBG] election_strategy: 1 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.358956+0000 mon.a (mon.0) 27 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.358956+0000 mon.a (mon.0) 27 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.358970+0000 mon.a (mon.0) 28 : cluster [DBG] 1: [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon.b 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.358970+0000 mon.a (mon.0) 28 : cluster [DBG] 1: [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon.b 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.359263+0000 mon.a (mon.0) 29 : cluster [DBG] fsmap 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.359263+0000 mon.a (mon.0) 29 : cluster [DBG] fsmap 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.359295+0000 mon.a (mon.0) 30 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.359295+0000 mon.a (mon.0) 30 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.359472+0000 mon.a (mon.0) 31 : cluster [DBG] mgrmap e14: a(active, since 3m), standbys: b 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.359472+0000 mon.a (mon.0) 31 : cluster [DBG] mgrmap e14: a(active, since 3m), standbys: b 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.359558+0000 mon.a (mon.0) 32 : cluster [INF] overall HEALTH_OK 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: cluster 2026-03-07T10:12:17.359558+0000 mon.a (mon.0) 32 : cluster [INF] overall HEALTH_OK 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: audit 2026-03-07T10:12:17.365266+0000 mon.a (mon.0) 33 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: audit 2026-03-07T10:12:17.365266+0000 mon.a (mon.0) 33 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: audit 2026-03-07T10:12:17.369686+0000 mon.a (mon.0) 34 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: audit 2026-03-07T10:12:17.369686+0000 mon.a (mon.0) 34 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: audit 2026-03-07T10:12:17.371353+0000 mon.a (mon.0) 35 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:12:18.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:18 vm10 bash[28189]: audit 2026-03-07T10:12:17.371353+0000 mon.a (mon.0) 35 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:12:18.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.352298+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-07T10:12:18.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.352298+0000 mon.b (mon.1) 1 : cluster [INF] mon.b calling monitor election 2026-03-07T10:12:18.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.352793+0000 mon.a (mon.0) 19 : cluster [INF] mon.a calling monitor election 2026-03-07T10:12:18.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.352793+0000 mon.a (mon.0) 19 : cluster [INF] mon.a calling monitor election 2026-03-07T10:12:18.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.354824+0000 mon.a (mon.0) 20 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-07T10:12:18.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.354824+0000 mon.a (mon.0) 20 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-07T10:12:18.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.358794+0000 mon.a (mon.0) 21 : cluster [DBG] monmap epoch 2 2026-03-07T10:12:18.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.358794+0000 mon.a (mon.0) 21 : cluster [DBG] monmap epoch 2 2026-03-07T10:12:18.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.358871+0000 mon.a (mon.0) 22 : cluster [DBG] fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:12:18.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.358871+0000 mon.a (mon.0) 22 : cluster [DBG] fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:12:18.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.358894+0000 mon.a (mon.0) 23 : cluster [DBG] last_changed 2026-03-07T10:09:14.324495+0000 2026-03-07T10:12:18.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.358894+0000 mon.a (mon.0) 23 : cluster [DBG] last_changed 2026-03-07T10:09:14.324495+0000 2026-03-07T10:12:18.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.358910+0000 mon.a (mon.0) 24 : cluster [DBG] created 2026-03-07T10:07:51.937005+0000 2026-03-07T10:12:18.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.358910+0000 mon.a (mon.0) 24 : cluster [DBG] created 2026-03-07T10:07:51.937005+0000 2026-03-07T10:12:18.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.358927+0000 mon.a (mon.0) 25 : cluster [DBG] min_mon_release 19 (squid) 2026-03-07T10:12:18.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.358927+0000 mon.a (mon.0) 25 : cluster [DBG] min_mon_release 19 (squid) 2026-03-07T10:12:18.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.358942+0000 mon.a (mon.0) 26 : cluster [DBG] election_strategy: 1 2026-03-07T10:12:18.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.358942+0000 mon.a (mon.0) 26 : cluster [DBG] election_strategy: 1 2026-03-07T10:12:18.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.358956+0000 mon.a (mon.0) 27 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-07T10:12:18.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.358956+0000 mon.a (mon.0) 27 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-07T10:12:18.713 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.358970+0000 mon.a (mon.0) 28 : cluster [DBG] 1: [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon.b 2026-03-07T10:12:18.713 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.358970+0000 mon.a (mon.0) 28 : cluster [DBG] 1: [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon.b 2026-03-07T10:12:18.713 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.359263+0000 mon.a (mon.0) 29 : cluster [DBG] fsmap 2026-03-07T10:12:18.713 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.359263+0000 mon.a (mon.0) 29 : cluster [DBG] fsmap 2026-03-07T10:12:18.713 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.359295+0000 mon.a (mon.0) 30 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-07T10:12:18.713 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.359295+0000 mon.a (mon.0) 30 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-07T10:12:18.713 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.359472+0000 mon.a (mon.0) 31 : cluster [DBG] mgrmap e14: a(active, since 3m), standbys: b 2026-03-07T10:12:18.713 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.359472+0000 mon.a (mon.0) 31 : cluster [DBG] mgrmap e14: a(active, since 3m), standbys: b 2026-03-07T10:12:18.713 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.359558+0000 mon.a (mon.0) 32 : cluster [INF] overall HEALTH_OK 2026-03-07T10:12:18.713 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: cluster 2026-03-07T10:12:17.359558+0000 mon.a (mon.0) 32 : cluster [INF] overall HEALTH_OK 2026-03-07T10:12:18.713 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: audit 2026-03-07T10:12:17.365266+0000 mon.a (mon.0) 33 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:18.713 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: audit 2026-03-07T10:12:17.365266+0000 mon.a (mon.0) 33 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:18.713 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: audit 2026-03-07T10:12:17.369686+0000 mon.a (mon.0) 34 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:18.713 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: audit 2026-03-07T10:12:17.369686+0000 mon.a (mon.0) 34 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:18.713 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: audit 2026-03-07T10:12:17.371353+0000 mon.a (mon.0) 35 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:12:18.713 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:18 vm02 bash[34866]: audit 2026-03-07T10:12:17.371353+0000 mon.a (mon.0) 35 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:12:19.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:19 vm10 bash[28189]: cluster 2026-03-07T10:12:18.292075+0000 mgr.a (mgr.14156) 173 : cluster [DBG] pgmap v115: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:19.696 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:19 vm10 bash[28189]: cluster 2026-03-07T10:12:18.292075+0000 mgr.a (mgr.14156) 173 : cluster [DBG] pgmap v115: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:19.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:19 vm02 bash[34866]: cluster 2026-03-07T10:12:18.292075+0000 mgr.a (mgr.14156) 173 : cluster [DBG] pgmap v115: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:19.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:19 vm02 bash[34866]: cluster 2026-03-07T10:12:18.292075+0000 mgr.a (mgr.14156) 173 : cluster [DBG] pgmap v115: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:20.717 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config 2026-03-07T10:12:21.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:21 vm10 bash[28189]: cluster 2026-03-07T10:12:20.292334+0000 mgr.a (mgr.14156) 174 : cluster [DBG] pgmap v116: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:21.695 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:21 vm10 bash[28189]: cluster 2026-03-07T10:12:20.292334+0000 mgr.a (mgr.14156) 174 : cluster [DBG] pgmap v116: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:21.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:21 vm02 bash[34866]: cluster 2026-03-07T10:12:20.292334+0000 mgr.a (mgr.14156) 174 : cluster [DBG] pgmap v116: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:21.712 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:21 vm02 bash[34866]: cluster 2026-03-07T10:12:20.292334+0000 mgr.a (mgr.14156) 174 : cluster [DBG] pgmap v116: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:23.446 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:23 vm10 bash[28189]: audit 2026-03-07T10:12:22.128176+0000 mon.a (mon.0) 36 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:23.447 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:23 vm10 bash[28189]: audit 2026-03-07T10:12:22.128176+0000 mon.a (mon.0) 36 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:23.447 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:23 vm10 bash[28189]: audit 2026-03-07T10:12:22.170193+0000 mon.a (mon.0) 37 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:23.447 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:23 vm10 bash[28189]: audit 2026-03-07T10:12:22.170193+0000 mon.a (mon.0) 37 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:23.447 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:23 vm10 bash[28189]: cluster 2026-03-07T10:12:22.292570+0000 mgr.a (mgr.14156) 175 : cluster [DBG] pgmap v117: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:23.447 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:23 vm10 bash[28189]: cluster 2026-03-07T10:12:22.292570+0000 mgr.a (mgr.14156) 175 : cluster [DBG] pgmap v117: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:23.448 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:23 vm10 bash[28189]: audit 2026-03-07T10:12:22.408176+0000 mon.a (mon.0) 38 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:23.448 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:23 vm10 bash[28189]: audit 2026-03-07T10:12:22.408176+0000 mon.a (mon.0) 38 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:23.448 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:23 vm10 bash[28189]: audit 2026-03-07T10:12:22.411185+0000 mon.a (mon.0) 39 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:23.448 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:23 vm10 bash[28189]: audit 2026-03-07T10:12:22.411185+0000 mon.a (mon.0) 39 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:23.448 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:23 vm10 bash[28189]: audit 2026-03-07T10:12:22.411787+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:12:23.448 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:23 vm10 bash[28189]: audit 2026-03-07T10:12:22.411787+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:12:23.448 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:23 vm10 bash[28189]: audit 2026-03-07T10:12:22.412301+0000 mon.a (mon.0) 41 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:12:23.448 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:23 vm10 bash[28189]: audit 2026-03-07T10:12:22.412301+0000 mon.a (mon.0) 41 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:12:23.448 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:23 vm10 bash[28189]: audit 2026-03-07T10:12:22.415229+0000 mon.a (mon.0) 42 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:23.448 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:23 vm10 bash[28189]: audit 2026-03-07T10:12:22.415229+0000 mon.a (mon.0) 42 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:23.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:23 vm02 bash[34866]: audit 2026-03-07T10:12:22.128176+0000 mon.a (mon.0) 36 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:23.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:23 vm02 bash[34866]: audit 2026-03-07T10:12:22.128176+0000 mon.a (mon.0) 36 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:23.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:23 vm02 bash[34866]: audit 2026-03-07T10:12:22.170193+0000 mon.a (mon.0) 37 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:23.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:23 vm02 bash[34866]: audit 2026-03-07T10:12:22.170193+0000 mon.a (mon.0) 37 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:23.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:23 vm02 bash[34866]: cluster 2026-03-07T10:12:22.292570+0000 mgr.a (mgr.14156) 175 : cluster [DBG] pgmap v117: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:23.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:23 vm02 bash[34866]: cluster 2026-03-07T10:12:22.292570+0000 mgr.a (mgr.14156) 175 : cluster [DBG] pgmap v117: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:23.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:23 vm02 bash[34866]: audit 2026-03-07T10:12:22.408176+0000 mon.a (mon.0) 38 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:23.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:23 vm02 bash[34866]: audit 2026-03-07T10:12:22.408176+0000 mon.a (mon.0) 38 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:23.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:23 vm02 bash[34866]: audit 2026-03-07T10:12:22.411185+0000 mon.a (mon.0) 39 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:23.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:23 vm02 bash[34866]: audit 2026-03-07T10:12:22.411185+0000 mon.a (mon.0) 39 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:23.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:23 vm02 bash[34866]: audit 2026-03-07T10:12:22.411787+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:12:23.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:23 vm02 bash[34866]: audit 2026-03-07T10:12:22.411787+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T10:12:23.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:23 vm02 bash[34866]: audit 2026-03-07T10:12:22.412301+0000 mon.a (mon.0) 41 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:12:23.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:23 vm02 bash[34866]: audit 2026-03-07T10:12:22.412301+0000 mon.a (mon.0) 41 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T10:12:23.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:23 vm02 bash[34866]: audit 2026-03-07T10:12:22.415229+0000 mon.a (mon.0) 42 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:23.462 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:23 vm02 bash[34866]: audit 2026-03-07T10:12:22.415229+0000 mon.a (mon.0) 42 : audit [INF] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' 2026-03-07T10:12:24.741 INFO:teuthology.orchestra.run.vm02.stderr:Using ceph image with id '8bccc98d839a' and tag 'cobaltcore-storage-v19.2.3-fasttrack-5' created on 2026-03-06 14:41:18 +0000 UTC 2026-03-07T10:12:24.741 INFO:teuthology.orchestra.run.vm02.stderr:harbor.clyso.com/custom-ceph/ceph/ceph@sha256:ffa52c72fad7bdd2657408de9cf8d87fc2c72f716d1a00277ba13f7c12b404e0 2026-03-07T10:12:25.658 INFO:teuthology.orchestra.run.vm02.stderr:+ FSID=6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:12:25.658 INFO:teuthology.orchestra.run.vm02.stderr:+ sleep 60 2026-03-07T10:12:25.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:25 vm10 bash[28189]: cluster 2026-03-07T10:12:24.292800+0000 mgr.a (mgr.14156) 176 : cluster [DBG] pgmap v118: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:25.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:25 vm10 bash[28189]: cluster 2026-03-07T10:12:24.292800+0000 mgr.a (mgr.14156) 176 : cluster [DBG] pgmap v118: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:25.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:25 vm02 bash[34866]: cluster 2026-03-07T10:12:24.292800+0000 mgr.a (mgr.14156) 176 : cluster [DBG] pgmap v118: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:25.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:25 vm02 bash[34866]: cluster 2026-03-07T10:12:24.292800+0000 mgr.a (mgr.14156) 176 : cluster [DBG] pgmap v118: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:26.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:26 vm10 bash[28189]: audit 2026-03-07T10:12:25.575689+0000 mon.a (mon.0) 43 : audit [DBG] from='client.? 192.168.123.102:0/1639059854' entity='client.admin' cmd=[{"prefix": "fsid"}]: dispatch 2026-03-07T10:12:26.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:26 vm10 bash[28189]: audit 2026-03-07T10:12:25.575689+0000 mon.a (mon.0) 43 : audit [DBG] from='client.? 192.168.123.102:0/1639059854' entity='client.admin' cmd=[{"prefix": "fsid"}]: dispatch 2026-03-07T10:12:26.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:26 vm02 bash[34866]: audit 2026-03-07T10:12:25.575689+0000 mon.a (mon.0) 43 : audit [DBG] from='client.? 192.168.123.102:0/1639059854' entity='client.admin' cmd=[{"prefix": "fsid"}]: dispatch 2026-03-07T10:12:26.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:26 vm02 bash[34866]: audit 2026-03-07T10:12:25.575689+0000 mon.a (mon.0) 43 : audit [DBG] from='client.? 192.168.123.102:0/1639059854' entity='client.admin' cmd=[{"prefix": "fsid"}]: dispatch 2026-03-07T10:12:27.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:27 vm10 bash[28189]: cluster 2026-03-07T10:12:26.292983+0000 mgr.a (mgr.14156) 177 : cluster [DBG] pgmap v119: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:27.946 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:27 vm10 bash[28189]: cluster 2026-03-07T10:12:26.292983+0000 mgr.a (mgr.14156) 177 : cluster [DBG] pgmap v119: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:27.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:27 vm02 bash[34866]: cluster 2026-03-07T10:12:26.292983+0000 mgr.a (mgr.14156) 177 : cluster [DBG] pgmap v119: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:27.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:27 vm02 bash[34866]: cluster 2026-03-07T10:12:26.292983+0000 mgr.a (mgr.14156) 177 : cluster [DBG] pgmap v119: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:29.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:29 vm10 bash[28189]: cluster 2026-03-07T10:12:28.293173+0000 mgr.a (mgr.14156) 178 : cluster [DBG] pgmap v120: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:29.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:29 vm10 bash[28189]: cluster 2026-03-07T10:12:28.293173+0000 mgr.a (mgr.14156) 178 : cluster [DBG] pgmap v120: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:29.961 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:29 vm02 bash[34866]: cluster 2026-03-07T10:12:28.293173+0000 mgr.a (mgr.14156) 178 : cluster [DBG] pgmap v120: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:29.961 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:29 vm02 bash[34866]: cluster 2026-03-07T10:12:28.293173+0000 mgr.a (mgr.14156) 178 : cluster [DBG] pgmap v120: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:31.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:31 vm10 bash[28189]: cluster 2026-03-07T10:12:30.293429+0000 mgr.a (mgr.14156) 179 : cluster [DBG] pgmap v121: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:31.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:31 vm10 bash[28189]: cluster 2026-03-07T10:12:30.293429+0000 mgr.a (mgr.14156) 179 : cluster [DBG] pgmap v121: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:31.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:31 vm02 bash[34866]: cluster 2026-03-07T10:12:30.293429+0000 mgr.a (mgr.14156) 179 : cluster [DBG] pgmap v121: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:31.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:31 vm02 bash[34866]: cluster 2026-03-07T10:12:30.293429+0000 mgr.a (mgr.14156) 179 : cluster [DBG] pgmap v121: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:33.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:33 vm10 bash[28189]: cluster 2026-03-07T10:12:32.293698+0000 mgr.a (mgr.14156) 180 : cluster [DBG] pgmap v122: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:33.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:33 vm10 bash[28189]: cluster 2026-03-07T10:12:32.293698+0000 mgr.a (mgr.14156) 180 : cluster [DBG] pgmap v122: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:33.961 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:33 vm02 bash[34866]: cluster 2026-03-07T10:12:32.293698+0000 mgr.a (mgr.14156) 180 : cluster [DBG] pgmap v122: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:33.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:33 vm02 bash[34866]: cluster 2026-03-07T10:12:32.293698+0000 mgr.a (mgr.14156) 180 : cluster [DBG] pgmap v122: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:35.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:35 vm10 bash[28189]: cluster 2026-03-07T10:12:34.293919+0000 mgr.a (mgr.14156) 181 : cluster [DBG] pgmap v123: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:35.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:35 vm10 bash[28189]: cluster 2026-03-07T10:12:34.293919+0000 mgr.a (mgr.14156) 181 : cluster [DBG] pgmap v123: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:35.961 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:35 vm02 bash[34866]: cluster 2026-03-07T10:12:34.293919+0000 mgr.a (mgr.14156) 181 : cluster [DBG] pgmap v123: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:35.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:35 vm02 bash[34866]: cluster 2026-03-07T10:12:34.293919+0000 mgr.a (mgr.14156) 181 : cluster [DBG] pgmap v123: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:37.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:37 vm10 bash[28189]: cluster 2026-03-07T10:12:36.294084+0000 mgr.a (mgr.14156) 182 : cluster [DBG] pgmap v124: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:37.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:37 vm10 bash[28189]: cluster 2026-03-07T10:12:36.294084+0000 mgr.a (mgr.14156) 182 : cluster [DBG] pgmap v124: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:37.961 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:37 vm02 bash[34866]: cluster 2026-03-07T10:12:36.294084+0000 mgr.a (mgr.14156) 182 : cluster [DBG] pgmap v124: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:37.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:37 vm02 bash[34866]: cluster 2026-03-07T10:12:36.294084+0000 mgr.a (mgr.14156) 182 : cluster [DBG] pgmap v124: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:39.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:39 vm10 bash[28189]: cluster 2026-03-07T10:12:38.294264+0000 mgr.a (mgr.14156) 183 : cluster [DBG] pgmap v125: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:39.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:39 vm10 bash[28189]: cluster 2026-03-07T10:12:38.294264+0000 mgr.a (mgr.14156) 183 : cluster [DBG] pgmap v125: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:39.961 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:39 vm02 bash[34866]: cluster 2026-03-07T10:12:38.294264+0000 mgr.a (mgr.14156) 183 : cluster [DBG] pgmap v125: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:39.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:39 vm02 bash[34866]: cluster 2026-03-07T10:12:38.294264+0000 mgr.a (mgr.14156) 183 : cluster [DBG] pgmap v125: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:41.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:41 vm10 bash[28189]: cluster 2026-03-07T10:12:40.294413+0000 mgr.a (mgr.14156) 184 : cluster [DBG] pgmap v126: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:41.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:41 vm10 bash[28189]: cluster 2026-03-07T10:12:40.294413+0000 mgr.a (mgr.14156) 184 : cluster [DBG] pgmap v126: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:41.961 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:41 vm02 bash[34866]: cluster 2026-03-07T10:12:40.294413+0000 mgr.a (mgr.14156) 184 : cluster [DBG] pgmap v126: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:41.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:41 vm02 bash[34866]: cluster 2026-03-07T10:12:40.294413+0000 mgr.a (mgr.14156) 184 : cluster [DBG] pgmap v126: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:43.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:43 vm10 bash[28189]: cluster 2026-03-07T10:12:42.294567+0000 mgr.a (mgr.14156) 185 : cluster [DBG] pgmap v127: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:43.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:43 vm10 bash[28189]: cluster 2026-03-07T10:12:42.294567+0000 mgr.a (mgr.14156) 185 : cluster [DBG] pgmap v127: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:43.961 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:43 vm02 bash[34866]: cluster 2026-03-07T10:12:42.294567+0000 mgr.a (mgr.14156) 185 : cluster [DBG] pgmap v127: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:43.961 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:43 vm02 bash[34866]: cluster 2026-03-07T10:12:42.294567+0000 mgr.a (mgr.14156) 185 : cluster [DBG] pgmap v127: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:45.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:45 vm10 bash[28189]: cluster 2026-03-07T10:12:44.294779+0000 mgr.a (mgr.14156) 186 : cluster [DBG] pgmap v128: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:45.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:45 vm10 bash[28189]: cluster 2026-03-07T10:12:44.294779+0000 mgr.a (mgr.14156) 186 : cluster [DBG] pgmap v128: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:45.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:45 vm02 bash[34866]: cluster 2026-03-07T10:12:44.294779+0000 mgr.a (mgr.14156) 186 : cluster [DBG] pgmap v128: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:45.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:45 vm02 bash[34866]: cluster 2026-03-07T10:12:44.294779+0000 mgr.a (mgr.14156) 186 : cluster [DBG] pgmap v128: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:47.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:47 vm10 bash[28189]: cluster 2026-03-07T10:12:46.294986+0000 mgr.a (mgr.14156) 187 : cluster [DBG] pgmap v129: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:47.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:47 vm10 bash[28189]: cluster 2026-03-07T10:12:46.294986+0000 mgr.a (mgr.14156) 187 : cluster [DBG] pgmap v129: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:47.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:47 vm02 bash[34866]: cluster 2026-03-07T10:12:46.294986+0000 mgr.a (mgr.14156) 187 : cluster [DBG] pgmap v129: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:47.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:47 vm02 bash[34866]: cluster 2026-03-07T10:12:46.294986+0000 mgr.a (mgr.14156) 187 : cluster [DBG] pgmap v129: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:49.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:49 vm10 bash[28189]: cluster 2026-03-07T10:12:48.295178+0000 mgr.a (mgr.14156) 188 : cluster [DBG] pgmap v130: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:49.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:49 vm10 bash[28189]: cluster 2026-03-07T10:12:48.295178+0000 mgr.a (mgr.14156) 188 : cluster [DBG] pgmap v130: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:49.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:49 vm02 bash[34866]: cluster 2026-03-07T10:12:48.295178+0000 mgr.a (mgr.14156) 188 : cluster [DBG] pgmap v130: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:49.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:49 vm02 bash[34866]: cluster 2026-03-07T10:12:48.295178+0000 mgr.a (mgr.14156) 188 : cluster [DBG] pgmap v130: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:51.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:51 vm10 bash[28189]: cluster 2026-03-07T10:12:50.295325+0000 mgr.a (mgr.14156) 189 : cluster [DBG] pgmap v131: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:51.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:51 vm10 bash[28189]: cluster 2026-03-07T10:12:50.295325+0000 mgr.a (mgr.14156) 189 : cluster [DBG] pgmap v131: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:51.961 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:51 vm02 bash[34866]: cluster 2026-03-07T10:12:50.295325+0000 mgr.a (mgr.14156) 189 : cluster [DBG] pgmap v131: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:51.962 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:51 vm02 bash[34866]: cluster 2026-03-07T10:12:50.295325+0000 mgr.a (mgr.14156) 189 : cluster [DBG] pgmap v131: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:53.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:53 vm10 bash[28189]: cluster 2026-03-07T10:12:52.295465+0000 mgr.a (mgr.14156) 190 : cluster [DBG] pgmap v132: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:53.945 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:53 vm10 bash[28189]: cluster 2026-03-07T10:12:52.295465+0000 mgr.a (mgr.14156) 190 : cluster [DBG] pgmap v132: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:53.961 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:53 vm02 bash[34866]: cluster 2026-03-07T10:12:52.295465+0000 mgr.a (mgr.14156) 190 : cluster [DBG] pgmap v132: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:53.961 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:53 vm02 bash[34866]: cluster 2026-03-07T10:12:52.295465+0000 mgr.a (mgr.14156) 190 : cluster [DBG] pgmap v132: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:56.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:55 vm10 bash[28189]: cluster 2026-03-07T10:12:54.295690+0000 mgr.a (mgr.14156) 191 : cluster [DBG] pgmap v133: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:56.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:55 vm10 bash[28189]: cluster 2026-03-07T10:12:54.295690+0000 mgr.a (mgr.14156) 191 : cluster [DBG] pgmap v133: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:56.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:55 vm02 bash[34866]: cluster 2026-03-07T10:12:54.295690+0000 mgr.a (mgr.14156) 191 : cluster [DBG] pgmap v133: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:56.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:55 vm02 bash[34866]: cluster 2026-03-07T10:12:54.295690+0000 mgr.a (mgr.14156) 191 : cluster [DBG] pgmap v133: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:58.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:57 vm10 bash[28189]: cluster 2026-03-07T10:12:56.295923+0000 mgr.a (mgr.14156) 192 : cluster [DBG] pgmap v134: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:58.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:57 vm10 bash[28189]: cluster 2026-03-07T10:12:56.295923+0000 mgr.a (mgr.14156) 192 : cluster [DBG] pgmap v134: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:58.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:57 vm02 bash[34866]: cluster 2026-03-07T10:12:56.295923+0000 mgr.a (mgr.14156) 192 : cluster [DBG] pgmap v134: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:12:58.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:57 vm02 bash[34866]: cluster 2026-03-07T10:12:56.295923+0000 mgr.a (mgr.14156) 192 : cluster [DBG] pgmap v134: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:00.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:59 vm10 bash[28189]: cluster 2026-03-07T10:12:58.296186+0000 mgr.a (mgr.14156) 193 : cluster [DBG] pgmap v135: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:00.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:12:59 vm10 bash[28189]: cluster 2026-03-07T10:12:58.296186+0000 mgr.a (mgr.14156) 193 : cluster [DBG] pgmap v135: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:00.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:59 vm02 bash[34866]: cluster 2026-03-07T10:12:58.296186+0000 mgr.a (mgr.14156) 193 : cluster [DBG] pgmap v135: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:00.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:12:59 vm02 bash[34866]: cluster 2026-03-07T10:12:58.296186+0000 mgr.a (mgr.14156) 193 : cluster [DBG] pgmap v135: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:02.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:01 vm10 bash[28189]: cluster 2026-03-07T10:13:00.296409+0000 mgr.a (mgr.14156) 194 : cluster [DBG] pgmap v136: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:02.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:01 vm10 bash[28189]: cluster 2026-03-07T10:13:00.296409+0000 mgr.a (mgr.14156) 194 : cluster [DBG] pgmap v136: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:02.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:01 vm02 bash[34866]: cluster 2026-03-07T10:13:00.296409+0000 mgr.a (mgr.14156) 194 : cluster [DBG] pgmap v136: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:02.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:01 vm02 bash[34866]: cluster 2026-03-07T10:13:00.296409+0000 mgr.a (mgr.14156) 194 : cluster [DBG] pgmap v136: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:04.114 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:03 vm02 bash[34866]: cluster 2026-03-07T10:13:02.296637+0000 mgr.a (mgr.14156) 195 : cluster [DBG] pgmap v137: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:04.114 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:03 vm02 bash[34866]: cluster 2026-03-07T10:13:02.296637+0000 mgr.a (mgr.14156) 195 : cluster [DBG] pgmap v137: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:04.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:03 vm10 bash[28189]: cluster 2026-03-07T10:13:02.296637+0000 mgr.a (mgr.14156) 195 : cluster [DBG] pgmap v137: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:04.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:03 vm10 bash[28189]: cluster 2026-03-07T10:13:02.296637+0000 mgr.a (mgr.14156) 195 : cluster [DBG] pgmap v137: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:06.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:05 vm10 bash[28189]: cluster 2026-03-07T10:13:04.296892+0000 mgr.a (mgr.14156) 196 : cluster [DBG] pgmap v138: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:06.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:05 vm10 bash[28189]: cluster 2026-03-07T10:13:04.296892+0000 mgr.a (mgr.14156) 196 : cluster [DBG] pgmap v138: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:06.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:05 vm02 bash[34866]: cluster 2026-03-07T10:13:04.296892+0000 mgr.a (mgr.14156) 196 : cluster [DBG] pgmap v138: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:06.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:05 vm02 bash[34866]: cluster 2026-03-07T10:13:04.296892+0000 mgr.a (mgr.14156) 196 : cluster [DBG] pgmap v138: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:08.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:07 vm10 bash[28189]: cluster 2026-03-07T10:13:06.297168+0000 mgr.a (mgr.14156) 197 : cluster [DBG] pgmap v139: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:08.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:07 vm10 bash[28189]: cluster 2026-03-07T10:13:06.297168+0000 mgr.a (mgr.14156) 197 : cluster [DBG] pgmap v139: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:08.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:07 vm02 bash[34866]: cluster 2026-03-07T10:13:06.297168+0000 mgr.a (mgr.14156) 197 : cluster [DBG] pgmap v139: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:08.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:07 vm02 bash[34866]: cluster 2026-03-07T10:13:06.297168+0000 mgr.a (mgr.14156) 197 : cluster [DBG] pgmap v139: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:10.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:09 vm10 bash[28189]: cluster 2026-03-07T10:13:08.297448+0000 mgr.a (mgr.14156) 198 : cluster [DBG] pgmap v140: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:10.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:09 vm10 bash[28189]: cluster 2026-03-07T10:13:08.297448+0000 mgr.a (mgr.14156) 198 : cluster [DBG] pgmap v140: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:10.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:09 vm02 bash[34866]: cluster 2026-03-07T10:13:08.297448+0000 mgr.a (mgr.14156) 198 : cluster [DBG] pgmap v140: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:10.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:09 vm02 bash[34866]: cluster 2026-03-07T10:13:08.297448+0000 mgr.a (mgr.14156) 198 : cluster [DBG] pgmap v140: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:12.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:11 vm10 bash[28189]: cluster 2026-03-07T10:13:10.297688+0000 mgr.a (mgr.14156) 199 : cluster [DBG] pgmap v141: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:12.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:11 vm10 bash[28189]: cluster 2026-03-07T10:13:10.297688+0000 mgr.a (mgr.14156) 199 : cluster [DBG] pgmap v141: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:12.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:11 vm02 bash[34866]: cluster 2026-03-07T10:13:10.297688+0000 mgr.a (mgr.14156) 199 : cluster [DBG] pgmap v141: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:12.212 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:11 vm02 bash[34866]: cluster 2026-03-07T10:13:10.297688+0000 mgr.a (mgr.14156) 199 : cluster [DBG] pgmap v141: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:14.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:13 vm10 bash[28189]: cluster 2026-03-07T10:13:12.297892+0000 mgr.a (mgr.14156) 200 : cluster [DBG] pgmap v142: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:14.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:13 vm10 bash[28189]: cluster 2026-03-07T10:13:12.297892+0000 mgr.a (mgr.14156) 200 : cluster [DBG] pgmap v142: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:14.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:13 vm02 bash[34866]: cluster 2026-03-07T10:13:12.297892+0000 mgr.a (mgr.14156) 200 : cluster [DBG] pgmap v142: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:14.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:13 vm02 bash[34866]: cluster 2026-03-07T10:13:12.297892+0000 mgr.a (mgr.14156) 200 : cluster [DBG] pgmap v142: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:16.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:15 vm10 bash[28189]: cluster 2026-03-07T10:13:14.298128+0000 mgr.a (mgr.14156) 201 : cluster [DBG] pgmap v143: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:16.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:15 vm10 bash[28189]: cluster 2026-03-07T10:13:14.298128+0000 mgr.a (mgr.14156) 201 : cluster [DBG] pgmap v143: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:16.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:15 vm02 bash[34866]: cluster 2026-03-07T10:13:14.298128+0000 mgr.a (mgr.14156) 201 : cluster [DBG] pgmap v143: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:16.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:15 vm02 bash[34866]: cluster 2026-03-07T10:13:14.298128+0000 mgr.a (mgr.14156) 201 : cluster [DBG] pgmap v143: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:18.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:17 vm10 bash[28189]: cluster 2026-03-07T10:13:16.298397+0000 mgr.a (mgr.14156) 202 : cluster [DBG] pgmap v144: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:18.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:17 vm10 bash[28189]: cluster 2026-03-07T10:13:16.298397+0000 mgr.a (mgr.14156) 202 : cluster [DBG] pgmap v144: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:18.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:17 vm02 bash[34866]: cluster 2026-03-07T10:13:16.298397+0000 mgr.a (mgr.14156) 202 : cluster [DBG] pgmap v144: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:18.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:17 vm02 bash[34866]: cluster 2026-03-07T10:13:16.298397+0000 mgr.a (mgr.14156) 202 : cluster [DBG] pgmap v144: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:20.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:19 vm10 bash[28189]: cluster 2026-03-07T10:13:18.298661+0000 mgr.a (mgr.14156) 203 : cluster [DBG] pgmap v145: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:20.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:19 vm10 bash[28189]: cluster 2026-03-07T10:13:18.298661+0000 mgr.a (mgr.14156) 203 : cluster [DBG] pgmap v145: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:20.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:19 vm02 bash[34866]: cluster 2026-03-07T10:13:18.298661+0000 mgr.a (mgr.14156) 203 : cluster [DBG] pgmap v145: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:20.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:19 vm02 bash[34866]: cluster 2026-03-07T10:13:18.298661+0000 mgr.a (mgr.14156) 203 : cluster [DBG] pgmap v145: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:22.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:21 vm10 bash[28189]: cluster 2026-03-07T10:13:20.298879+0000 mgr.a (mgr.14156) 204 : cluster [DBG] pgmap v146: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:22.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:21 vm10 bash[28189]: cluster 2026-03-07T10:13:20.298879+0000 mgr.a (mgr.14156) 204 : cluster [DBG] pgmap v146: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:22.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:21 vm02 bash[34866]: cluster 2026-03-07T10:13:20.298879+0000 mgr.a (mgr.14156) 204 : cluster [DBG] pgmap v146: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:22.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:21 vm02 bash[34866]: cluster 2026-03-07T10:13:20.298879+0000 mgr.a (mgr.14156) 204 : cluster [DBG] pgmap v146: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:23.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:22 vm10 bash[28189]: audit 2026-03-07T10:13:22.425325+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:13:23.196 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:22 vm10 bash[28189]: audit 2026-03-07T10:13:22.425325+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:13:23.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:22 vm02 bash[34866]: audit 2026-03-07T10:13:22.425325+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:13:23.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:22 vm02 bash[34866]: audit 2026-03-07T10:13:22.425325+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14156 192.168.123.102:0/1666535981' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T10:13:24.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:23 vm10 bash[28189]: cluster 2026-03-07T10:13:22.299049+0000 mgr.a (mgr.14156) 205 : cluster [DBG] pgmap v147: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:24.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:23 vm10 bash[28189]: cluster 2026-03-07T10:13:22.299049+0000 mgr.a (mgr.14156) 205 : cluster [DBG] pgmap v147: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:24.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:23 vm02 bash[34866]: cluster 2026-03-07T10:13:22.299049+0000 mgr.a (mgr.14156) 205 : cluster [DBG] pgmap v147: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:24.211 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:23 vm02 bash[34866]: cluster 2026-03-07T10:13:22.299049+0000 mgr.a (mgr.14156) 205 : cluster [DBG] pgmap v147: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:25.658 INFO:teuthology.orchestra.run.vm02.stderr:+ grep '\-\-cpus=2' /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/unit.run 2026-03-07T10:13:25.659 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-mon --privileged --group-add=disk --init --name ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e-mon-a --pids-limit=0 --cpus=2 -e CONTAINER_IMAGE=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:ffa52c72fad7bdd2657408de9cf8d87fc2c72f716d1a00277ba13f7c12b404e0 -e NODE_NAME=vm02 -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e:/var/run/ceph:z -v /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e:/var/log/ceph:z -v /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/crash:/var/lib/ceph/crash:z -v /dev:/dev -v /run/udev:/run/udev -v /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a:/var/lib/ceph/mon/ceph-a:z -v /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config:/etc/ceph/ceph.conf:z harbor.clyso.com/custom-ceph/ceph/ceph@sha256:ffa52c72fad7bdd2657408de9cf8d87fc2c72f716d1a00277ba13f7c12b404e0 -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true '--default-log-stderr-prefix=debug ' --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true --debug_ms 10 2026-03-07T10:13:25.659 INFO:teuthology.orchestra.run.vm02.stderr:+ grep '\-\-debug_ms 10' /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/unit.run 2026-03-07T10:13:25.659 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-mon --privileged --group-add=disk --init --name ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e-mon-a --pids-limit=0 --cpus=2 -e CONTAINER_IMAGE=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:ffa52c72fad7bdd2657408de9cf8d87fc2c72f716d1a00277ba13f7c12b404e0 -e NODE_NAME=vm02 -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e:/var/run/ceph:z -v /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e:/var/log/ceph:z -v /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/crash:/var/lib/ceph/crash:z -v /dev:/dev -v /run/udev:/run/udev -v /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a:/var/lib/ceph/mon/ceph-a:z -v /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/mon.a/config:/etc/ceph/ceph.conf:z harbor.clyso.com/custom-ceph/ceph/ceph@sha256:ffa52c72fad7bdd2657408de9cf8d87fc2c72f716d1a00277ba13f7c12b404e0 -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true '--default-log-stderr-prefix=debug ' --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true --debug_ms 10 2026-03-07T10:13:25.660 INFO:teuthology.orchestra.run.vm02.stderr:+ grep testing_custom_containers /etc/cephadm_testing/testing.txt 2026-03-07T10:13:25.660 INFO:teuthology.orchestra.run.vm02.stdout:testing_custom_containers 2026-03-07T10:13:25.660 INFO:teuthology.orchestra.run.vm02.stderr:++ find /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e -maxdepth 1 -type d -name 'container.bar.*' 2026-03-07T10:13:25.661 INFO:teuthology.orchestra.run.vm02.stderr:+ dir=/var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/container.bar.vm02 2026-03-07T10:13:25.661 INFO:teuthology.orchestra.run.vm02.stderr:+ test -n /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/container.bar.vm02 2026-03-07T10:13:25.661 INFO:teuthology.orchestra.run.vm02.stderr:+ grep ok /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/container.bar.vm02/data/primary.txt 2026-03-07T10:13:25.662 INFO:teuthology.orchestra.run.vm02.stdout:ok 2026-03-07T10:13:25.662 INFO:teuthology.orchestra.run.vm02.stderr:+ grep from=centos /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/container.bar.vm02/data/from.txt 2026-03-07T10:13:25.663 INFO:teuthology.orchestra.run.vm02.stdout:from=centos 2026-03-07T10:13:25.663 INFO:teuthology.orchestra.run.vm02.stderr:+ test -s /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/container.bar.vm02/data/presized.dat 2026-03-07T10:13:25.664 DEBUG:teuthology.run_tasks:Unwinding manager cephadm 2026-03-07T10:13:25.666 INFO:tasks.cephadm:Teardown begin 2026-03-07T10:13:25.666 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-07T10:13:25.714 DEBUG:teuthology.orchestra.run.vm10:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-07T10:13:25.722 INFO:tasks.cephadm:Cleaning up testdir ceph.* files... 2026-03-07T10:13:25.722 DEBUG:teuthology.orchestra.run.vm02:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-07T10:13:25.758 DEBUG:teuthology.orchestra.run.vm10:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-07T10:13:25.766 INFO:tasks.cephadm:Stopping all daemons... 2026-03-07T10:13:25.766 INFO:tasks.cephadm.mon.a:Stopping mon.a... 2026-03-07T10:13:25.766 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@mon.a 2026-03-07T10:13:25.914 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:25 vm02 systemd[1]: Stopping Ceph mon.a for 6c715b7a-1a0d-11f1-b180-89615ccd948e... 2026-03-07T10:13:25.914 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:25 vm02 bash[34866]: cluster 2026-03-07T10:13:24.299235+0000 mgr.a (mgr.14156) 206 : cluster [DBG] pgmap v148: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:25.914 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:25 vm02 bash[34866]: cluster 2026-03-07T10:13:24.299235+0000 mgr.a (mgr.14156) 206 : cluster [DBG] pgmap v148: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:25.914 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:25 vm02 bash[34866]: debug 2026-03-07T10:13:25.842+0000 7f955ab81640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true --debug_ms 10 (PID: 1) UID: 0 2026-03-07T10:13:25.914 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:25 vm02 bash[34866]: debug 2026-03-07T10:13:25.842+0000 7f955ab81640 -1 mon.a@0(leader) e2 *** Got Signal Terminated *** 2026-03-07T10:13:25.914 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 07 10:13:25 vm02 bash[35277]: ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e-mon-a 2026-03-07T10:13:25.947 DEBUG:teuthology.orchestra.run.vm02:> sudo pkill -f 'journalctl -f -n 0 -u ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@mon.a.service' 2026-03-07T10:13:25.958 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-07T10:13:25.958 INFO:tasks.cephadm.mon.a:Stopped mon.a 2026-03-07T10:13:25.958 INFO:tasks.cephadm.mon.b:Stopping mon.b... 2026-03-07T10:13:25.958 DEBUG:teuthology.orchestra.run.vm10:> sudo systemctl stop ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@mon.b 2026-03-07T10:13:26.007 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:25 vm10 bash[28189]: cluster 2026-03-07T10:13:24.299235+0000 mgr.a (mgr.14156) 206 : cluster [DBG] pgmap v148: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:26.007 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:25 vm10 bash[28189]: cluster 2026-03-07T10:13:24.299235+0000 mgr.a (mgr.14156) 206 : cluster [DBG] pgmap v148: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-07T10:13:26.007 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:25 vm10 systemd[1]: Stopping Ceph mon.b for 6c715b7a-1a0d-11f1-b180-89615ccd948e... 2026-03-07T10:13:26.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:26 vm10 bash[28189]: debug 2026-03-07T10:13:26.004+0000 7f9532560640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.b -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true --debug_ms 10 (PID: 1) UID: 0 2026-03-07T10:13:26.195 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 07 10:13:26 vm10 bash[28189]: debug 2026-03-07T10:13:26.004+0000 7f9532560640 -1 mon.b@1(peon) e2 *** Got Signal Terminated *** 2026-03-07T10:13:26.262 DEBUG:teuthology.orchestra.run.vm10:> sudo pkill -f 'journalctl -f -n 0 -u ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@mon.b.service' 2026-03-07T10:13:26.272 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-07T10:13:26.272 INFO:tasks.cephadm.mon.b:Stopped mon.b 2026-03-07T10:13:26.272 INFO:tasks.cephadm.mgr.a:Stopping mgr.a... 2026-03-07T10:13:26.272 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@mgr.a 2026-03-07T10:13:26.413 DEBUG:teuthology.orchestra.run.vm02:> sudo pkill -f 'journalctl -f -n 0 -u ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@mgr.a.service' 2026-03-07T10:13:26.422 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-07T10:13:26.422 INFO:tasks.cephadm.mgr.a:Stopped mgr.a 2026-03-07T10:13:26.423 INFO:tasks.cephadm.mgr.b:Stopping mgr.b... 2026-03-07T10:13:26.423 DEBUG:teuthology.orchestra.run.vm10:> sudo systemctl stop ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@mgr.b 2026-03-07T10:13:26.489 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 07 10:13:26 vm10 systemd[1]: Stopping Ceph mgr.b for 6c715b7a-1a0d-11f1-b180-89615ccd948e... 2026-03-07T10:13:26.555 DEBUG:teuthology.orchestra.run.vm10:> sudo pkill -f 'journalctl -f -n 0 -u ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@mgr.b.service' 2026-03-07T10:13:26.566 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-07T10:13:26.566 INFO:tasks.cephadm.mgr.b:Stopped mgr.b 2026-03-07T10:13:26.566 INFO:tasks.cephadm.osd.0:Stopping osd.0... 2026-03-07T10:13:26.566 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@osd.0 2026-03-07T10:13:26.961 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 07 10:13:26 vm02 systemd[1]: Stopping Ceph osd.0 for 6c715b7a-1a0d-11f1-b180-89615ccd948e... 2026-03-07T10:13:26.961 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 07 10:13:26 vm02 bash[27327]: debug 2026-03-07T10:13:26.606+0000 7f093e83c640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug (PID: 1) UID: 0 2026-03-07T10:13:26.961 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 07 10:13:26 vm02 bash[27327]: debug 2026-03-07T10:13:26.606+0000 7f093e83c640 -1 osd.0 14 *** Got signal Terminated *** 2026-03-07T10:13:26.961 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 07 10:13:26 vm02 bash[27327]: debug 2026-03-07T10:13:26.606+0000 7f093e83c640 -1 osd.0 14 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-07T10:13:31.961 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 07 10:13:31 vm02 bash[35453]: ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e-osd-0 2026-03-07T10:13:32.264 DEBUG:teuthology.orchestra.run.vm02:> sudo pkill -f 'journalctl -f -n 0 -u ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@osd.0.service' 2026-03-07T10:13:32.288 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-07T10:13:32.288 INFO:tasks.cephadm.osd.0:Stopped osd.0 2026-03-07T10:13:32.288 INFO:tasks.cephadm.osd.1:Stopping osd.1... 2026-03-07T10:13:32.289 DEBUG:teuthology.orchestra.run.vm10:> sudo systemctl stop ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@osd.1 2026-03-07T10:13:32.695 INFO:journalctl@ceph.osd.1.vm10.stdout:Mar 07 10:13:32 vm10 systemd[1]: Stopping Ceph osd.1 for 6c715b7a-1a0d-11f1-b180-89615ccd948e... 2026-03-07T10:13:32.695 INFO:journalctl@ceph.osd.1.vm10.stdout:Mar 07 10:13:32 vm10 bash[22515]: debug 2026-03-07T10:13:32.332+0000 7fbadadb3640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug (PID: 1) UID: 0 2026-03-07T10:13:32.695 INFO:journalctl@ceph.osd.1.vm10.stdout:Mar 07 10:13:32 vm10 bash[22515]: debug 2026-03-07T10:13:32.332+0000 7fbadadb3640 -1 osd.1 14 *** Got signal Terminated *** 2026-03-07T10:13:32.696 INFO:journalctl@ceph.osd.1.vm10.stdout:Mar 07 10:13:32 vm10 bash[22515]: debug 2026-03-07T10:13:32.332+0000 7fbadadb3640 -1 osd.1 14 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-07T10:13:37.695 INFO:journalctl@ceph.osd.1.vm10.stdout:Mar 07 10:13:37 vm10 bash[28770]: ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e-osd-1 2026-03-07T10:13:38.358 DEBUG:teuthology.orchestra.run.vm10:> sudo pkill -f 'journalctl -f -n 0 -u ceph-6c715b7a-1a0d-11f1-b180-89615ccd948e@osd.1.service' 2026-03-07T10:13:38.381 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-07T10:13:38.381 INFO:tasks.cephadm.osd.1:Stopped osd.1 2026-03-07T10:13:38.381 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e --force --keep-logs 2026-03-07T10:13:38.617 INFO:teuthology.orchestra.run.vm02.stdout:Deleting cluster with fsid: 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:13:46.680 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e --force --keep-logs 2026-03-07T10:13:46.920 INFO:teuthology.orchestra.run.vm10.stdout:Deleting cluster with fsid: 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:13:53.458 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-07T10:13:53.465 DEBUG:teuthology.orchestra.run.vm10:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-07T10:13:53.472 INFO:tasks.cephadm:Archiving crash dumps... 2026-03-07T10:13:53.472 DEBUG:teuthology.misc:Transferring archived files from vm02:/var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/crash to /archive/irq0-2026-03-07_10:02:54-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps/9/remote/vm02/crash 2026-03-07T10:13:53.472 DEBUG:teuthology.orchestra.run.vm02:> sudo tar c -f - -C /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/crash -- . 2026-03-07T10:13:53.515 INFO:teuthology.orchestra.run.vm02.stderr:tar: /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/crash: Cannot open: No such file or directory 2026-03-07T10:13:53.515 INFO:teuthology.orchestra.run.vm02.stderr:tar: Error is not recoverable: exiting now 2026-03-07T10:13:53.515 DEBUG:teuthology.misc:Transferring archived files from vm10:/var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/crash to /archive/irq0-2026-03-07_10:02:54-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps/9/remote/vm10/crash 2026-03-07T10:13:53.515 DEBUG:teuthology.orchestra.run.vm10:> sudo tar c -f - -C /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/crash -- . 2026-03-07T10:13:53.522 INFO:teuthology.orchestra.run.vm10.stderr:tar: /var/lib/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/crash: Cannot open: No such file or directory 2026-03-07T10:13:53.522 INFO:teuthology.orchestra.run.vm10.stderr:tar: Error is not recoverable: exiting now 2026-03-07T10:13:53.523 INFO:tasks.cephadm:Checking cluster log for badness... 2026-03-07T10:13:53.523 DEBUG:teuthology.orchestra.run.vm02:> sudo egrep '\[ERR\]|\[WRN\]|\[SEC\]' /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph.log | egrep CEPHADM_ | egrep -v '\(MDS_ALL_DOWN\)' | egrep -v '\(MDS_UP_LESS_THAN_MAX\)' | egrep -v CEPHADM_FAILED_DAEMON | head -n 1 2026-03-07T10:13:53.569 INFO:tasks.cephadm:Compressing logs... 2026-03-07T10:13:53.569 DEBUG:teuthology.orchestra.run.vm02:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-07T10:13:53.612 DEBUG:teuthology.orchestra.run.vm10:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-07T10:13:53.618 INFO:teuthology.orchestra.run.vm10.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-07T10:13:53.618 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-07T10:13:53.618 INFO:teuthology.orchestra.run.vm02.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-07T10:13:53.619 INFO:teuthology.orchestra.run.vm10.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-07T10:13:53.619 INFO:teuthology.orchestra.run.vm10.stderr:gzip -5 --verbose -- /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph.log 2026-03-07T10:13:53.619 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph-mgr.a.log 2026-03-07T10:13:53.619 INFO:teuthology.orchestra.run.vm10.stderr:/var/log/ceph/cephadm.log: gzip -5 --verbose -- /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph-mon.b.log 2026-03-07T10:13:53.620 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph.log 2026-03-07T10:13:53.620 INFO:teuthology.orchestra.run.vm10.stderr:/var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph.log: 87.3% -- replaced with /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph.log.gz 2026-03-07T10:13:53.620 INFO:teuthology.orchestra.run.vm10.stderr: 88.6% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-07T10:13:53.620 INFO:teuthology.orchestra.run.vm10.stderr:gzip -5 --verbose -- /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph-osd.1.log 2026-03-07T10:13:53.620 INFO:teuthology.orchestra.run.vm10.stderr:gzip -5 --verbose -- /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph-mgr.b.log 2026-03-07T10:13:53.621 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/cephadm.log: /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph-mgr.a.log: 89.7% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-07T10:13:53.622 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph-mon.a.log 2026-03-07T10:13:53.622 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph.log: 87.1% -- replaced with /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph.log.gz 2026-03-07T10:13:53.623 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph.audit.log 2026-03-07T10:13:53.627 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph-mon.a.log: gzip -5 --verbose -- /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph-volume.log 2026-03-07T10:13:53.628 INFO:teuthology.orchestra.run.vm10.stderr:/var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph-mon.b.log: /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph-osd.1.log: gzip -5 --verbose -- /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph.audit.log 2026-03-07T10:13:53.630 INFO:teuthology.orchestra.run.vm10.stderr:/var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph-mgr.b.log: 91.2% -- replaced with /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph-mgr.b.log.gz 2026-03-07T10:13:53.631 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph.audit.log: 89.5% -- replaced with /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph.audit.log.gz 2026-03-07T10:13:53.633 INFO:teuthology.orchestra.run.vm10.stderr:gzip -5 --verbose -- /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph-volume.log 2026-03-07T10:13:53.633 INFO:teuthology.orchestra.run.vm10.stderr:/var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph.audit.log: 90.0% -- replaced with /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph.audit.log.gz 2026-03-07T10:13:53.634 INFO:teuthology.orchestra.run.vm10.stderr: 94.0% -- replaced with /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph-osd.1.log.gz 2026-03-07T10:13:53.634 INFO:teuthology.orchestra.run.vm10.stderr:gzip -5 --verbose -- /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph.cephadm.log 2026-03-07T10:13:53.639 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph.cephadm.log 2026-03-07T10:13:53.640 INFO:teuthology.orchestra.run.vm10.stderr:/var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph-volume.log: /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph.cephadm.log: 76.7% -- replaced with /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph.cephadm.log.gz 2026-03-07T10:13:53.640 INFO:teuthology.orchestra.run.vm10.stderr: 93.7% -- replaced with /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph-volume.log.gz 2026-03-07T10:13:53.645 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph-volume.log: gzip -5 --verbose -- /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph-osd.0.log 2026-03-07T10:13:53.645 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph.cephadm.log: 79.1% -- replaced with /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph.cephadm.log.gz 2026-03-07T10:13:53.645 INFO:teuthology.orchestra.run.vm02.stderr: 93.7% -- replaced with /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph-volume.log.gz 2026-03-07T10:13:53.652 INFO:teuthology.orchestra.run.vm02.stderr: 89.7% -- replaced with /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph-mgr.a.log.gz 2026-03-07T10:13:53.661 INFO:teuthology.orchestra.run.vm10.stderr: 93.4% -- replaced with /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph-mon.b.log.gz 2026-03-07T10:13:53.662 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph-osd.0.log: 94.1% -- replaced with /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph-osd.0.log.gz 2026-03-07T10:13:53.662 INFO:teuthology.orchestra.run.vm10.stderr: 2026-03-07T10:13:53.662 INFO:teuthology.orchestra.run.vm10.stderr:real 0m0.048s 2026-03-07T10:13:53.662 INFO:teuthology.orchestra.run.vm10.stderr:user 0m0.063s 2026-03-07T10:13:53.662 INFO:teuthology.orchestra.run.vm10.stderr:sys 0m0.008s 2026-03-07T10:13:53.730 INFO:teuthology.orchestra.run.vm02.stderr: 91.7% -- replaced with /var/log/ceph/6c715b7a-1a0d-11f1-b180-89615ccd948e/ceph-mon.a.log.gz 2026-03-07T10:13:53.732 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-07T10:13:53.732 INFO:teuthology.orchestra.run.vm02.stderr:real 0m0.118s 2026-03-07T10:13:53.732 INFO:teuthology.orchestra.run.vm02.stderr:user 0m0.156s 2026-03-07T10:13:53.732 INFO:teuthology.orchestra.run.vm02.stderr:sys 0m0.006s 2026-03-07T10:13:53.732 INFO:tasks.cephadm:Archiving logs... 2026-03-07T10:13:53.732 DEBUG:teuthology.misc:Transferring archived files from vm02:/var/log/ceph to /archive/irq0-2026-03-07_10:02:54-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps/9/remote/vm02/log 2026-03-07T10:13:53.732 DEBUG:teuthology.orchestra.run.vm02:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-07T10:13:53.799 DEBUG:teuthology.misc:Transferring archived files from vm10:/var/log/ceph to /archive/irq0-2026-03-07_10:02:54-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps/9/remote/vm10/log 2026-03-07T10:13:53.799 DEBUG:teuthology.orchestra.run.vm10:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-07T10:13:53.810 INFO:tasks.cephadm:Removing cluster... 2026-03-07T10:13:53.810 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e --force 2026-03-07T10:13:54.063 INFO:teuthology.orchestra.run.vm02.stdout:Deleting cluster with fsid: 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:13:55.135 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 6c715b7a-1a0d-11f1-b180-89615ccd948e --force 2026-03-07T10:13:55.363 INFO:teuthology.orchestra.run.vm10.stdout:Deleting cluster with fsid: 6c715b7a-1a0d-11f1-b180-89615ccd948e 2026-03-07T10:13:56.423 INFO:tasks.cephadm:Removing cephadm ... 2026-03-07T10:13:56.423 DEBUG:teuthology.orchestra.run.vm02:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-07T10:13:56.427 DEBUG:teuthology.orchestra.run.vm10:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-07T10:13:56.430 INFO:tasks.cephadm:Teardown complete 2026-03-07T10:13:56.430 DEBUG:teuthology.run_tasks:Unwinding manager clock 2026-03-07T10:13:56.432 INFO:teuthology.task.clock:Checking final clock skew... 2026-03-07T10:13:56.432 DEBUG:teuthology.orchestra.run.vm02:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-07T10:13:56.472 DEBUG:teuthology.orchestra.run.vm10:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-07T10:13:57.185 INFO:teuthology.orchestra.run.vm02.stdout: remote refid st t when poll reach delay offset jitter 2026-03-07T10:13:57.185 INFO:teuthology.orchestra.run.vm02.stdout:============================================================================== 2026-03-07T10:13:57.185 INFO:teuthology.orchestra.run.vm02.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T10:13:57.185 INFO:teuthology.orchestra.run.vm02.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T10:13:57.185 INFO:teuthology.orchestra.run.vm02.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T10:13:57.185 INFO:teuthology.orchestra.run.vm02.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T10:13:57.185 INFO:teuthology.orchestra.run.vm02.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T10:13:57.185 INFO:teuthology.orchestra.run.vm02.stdout:-ntp2.kernfusion 192.53.103.108 2 u 56 64 77 33.369 +1.241 1.204 2026-03-07T10:13:57.185 INFO:teuthology.orchestra.run.vm02.stdout:*130.61.89.107 237.17.204.95 2 u 52 64 77 20.953 +1.234 0.202 2026-03-07T10:13:57.185 INFO:teuthology.orchestra.run.vm02.stdout:-vps-fra8.orlean 195.145.119.188 2 u 55 64 77 30.491 -0.757 0.389 2026-03-07T10:13:57.185 INFO:teuthology.orchestra.run.vm02.stdout:-static.179.181. 161.62.157.173 3 u 55 64 77 23.610 +1.557 0.146 2026-03-07T10:13:57.185 INFO:teuthology.orchestra.run.vm02.stdout:-static.46.170.2 188.40.142.18 3 u 52 64 77 25.596 +0.732 0.321 2026-03-07T10:13:57.185 INFO:teuthology.orchestra.run.vm02.stdout:-time.ndless.net 192.53.103.108 2 u 55 64 77 28.846 +1.955 0.265 2026-03-07T10:13:57.185 INFO:teuthology.orchestra.run.vm02.stdout:-time.cloudflare 10.71.2.234 3 u 58 64 77 20.442 +2.410 0.232 2026-03-07T10:13:57.185 INFO:teuthology.orchestra.run.vm02.stdout:#172-104-149-161 80.192.165.246 2 u 51 64 77 22.536 -4.283 0.779 2026-03-07T10:13:57.185 INFO:teuthology.orchestra.run.vm02.stdout:+ntp1.adminforge 131.188.3.220 2 u 58 64 77 25.122 +1.057 0.178 2026-03-07T10:13:57.186 INFO:teuthology.orchestra.run.vm02.stdout:#netcup02.therav 189.97.54.122 2 u 54 64 77 28.386 -1.408 0.194 2026-03-07T10:13:57.186 INFO:teuthology.orchestra.run.vm02.stdout:+static.buzo.eu 100.10.69.89 2 u 51 64 77 23.582 +1.316 0.227 2026-03-07T10:13:57.186 INFO:teuthology.orchestra.run.vm02.stdout:-ntp2.uni-ulm.de 129.69.253.1 2 u 53 64 77 27.285 +0.142 0.241 2026-03-07T10:13:57.186 INFO:teuthology.orchestra.run.vm02.stdout:#128.127.67.142 189.97.54.122 2 u 59 64 73 30.003 +0.167 0.244 2026-03-07T10:13:57.186 INFO:teuthology.orchestra.run.vm02.stdout:#185.125.190.58 37.15.221.189 2 u 63 64 77 35.702 -0.179 0.383 2026-03-07T10:13:57.186 INFO:teuthology.orchestra.run.vm02.stdout:#funky.f5s.de 131.188.3.222 2 u 53 64 77 25.097 +1.204 0.569 2026-03-07T10:13:57.186 INFO:teuthology.orchestra.run.vm02.stdout:#netcup01.therav 171.237.1.87 2 u 48 64 77 28.325 -1.556 0.121 2026-03-07T10:13:57.186 INFO:teuthology.orchestra.run.vm02.stdout:#185.125.190.56 194.121.207.249 2 u 60 64 77 34.484 -0.521 0.167 2026-03-07T10:13:57.186 INFO:teuthology.orchestra.run.vm10.stdout: remote refid st t when poll reach delay offset jitter 2026-03-07T10:13:57.186 INFO:teuthology.orchestra.run.vm10.stdout:============================================================================== 2026-03-07T10:13:57.186 INFO:teuthology.orchestra.run.vm10.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T10:13:57.186 INFO:teuthology.orchestra.run.vm10.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T10:13:57.186 INFO:teuthology.orchestra.run.vm10.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T10:13:57.186 INFO:teuthology.orchestra.run.vm10.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T10:13:57.187 INFO:teuthology.orchestra.run.vm10.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T10:13:57.187 INFO:teuthology.orchestra.run.vm10.stdout:+time.ndless.net 192.53.103.108 2 u 64 64 77 29.150 +0.691 2.220 2026-03-07T10:13:57.187 INFO:teuthology.orchestra.run.vm10.stdout:*130.61.89.107 237.17.204.95 2 u 57 64 77 20.960 +0.001 2.076 2026-03-07T10:13:57.187 INFO:teuthology.orchestra.run.vm10.stdout:+static.179.181. 161.62.157.173 3 u 52 64 77 23.560 +0.592 0.846 2026-03-07T10:13:57.187 INFO:teuthology.orchestra.run.vm10.stdout:+ntp1.adminforge 131.188.3.220 2 u 56 64 77 25.008 +0.471 1.041 2026-03-07T10:13:57.187 INFO:teuthology.orchestra.run.vm10.stdout:+static.46.170.2 188.40.142.18 3 u 59 64 77 25.138 -0.305 1.801 2026-03-07T10:13:57.187 INFO:teuthology.orchestra.run.vm10.stdout:+ntp2.uni-ulm.de 129.69.253.1 2 u 57 64 77 27.414 +2.682 3.275 2026-03-07T10:13:57.187 INFO:teuthology.orchestra.run.vm10.stdout:#172-104-149-161 80.192.165.246 2 u 53 64 77 22.589 -6.325 0.982 2026-03-07T10:13:57.187 INFO:teuthology.orchestra.run.vm10.stdout:+funky.f5s.de 131.188.3.222 2 u 51 64 77 25.002 +0.084 0.843 2026-03-07T10:13:57.187 INFO:teuthology.orchestra.run.vm10.stdout:#128.127.67.142 189.97.54.122 2 u 52 64 77 29.882 -0.887 0.450 2026-03-07T10:13:57.187 INFO:teuthology.orchestra.run.vm10.stdout:+ntp2.kernfusion 192.53.103.108 2 u 56 64 77 32.801 +1.006 2.384 2026-03-07T10:13:57.187 INFO:teuthology.orchestra.run.vm10.stdout:#static.119.109. 131.188.3.223 2 u 55 64 77 23.531 +2.213 2.073 2026-03-07T10:13:57.187 INFO:teuthology.orchestra.run.vm10.stdout:+static.buzo.eu 100.10.69.89 2 u 51 64 77 23.560 -0.163 0.550 2026-03-07T10:13:57.187 INFO:teuthology.orchestra.run.vm10.stdout:+vps-fra8.orlean 195.145.119.188 2 u 55 64 77 32.408 -0.196 0.918 2026-03-07T10:13:57.187 INFO:teuthology.orchestra.run.vm10.stdout:#185.125.190.57 194.121.207.249 2 u 62 64 77 48.015 -8.475 5.506 2026-03-07T10:13:57.187 INFO:teuthology.orchestra.run.vm10.stdout:+time.cloudflare 10.214.8.5 3 u 56 64 77 20.371 +0.826 0.801 2026-03-07T10:13:57.187 INFO:teuthology.orchestra.run.vm10.stdout:#netcup01.therav 171.237.1.87 2 u 55 64 77 28.245 -2.929 0.356 2026-03-07T10:13:57.187 INFO:teuthology.orchestra.run.vm10.stdout:#185.125.190.56 194.121.207.249 2 u 1 64 177 34.351 +1.543 3.310 2026-03-07T10:13:57.187 DEBUG:teuthology.run_tasks:Unwinding manager ansible.cephlab 2026-03-07T10:13:57.190 INFO:teuthology.task.ansible:Skipping ansible cleanup... 2026-03-07T10:13:57.190 DEBUG:teuthology.run_tasks:Unwinding manager selinux 2026-03-07T10:13:57.193 DEBUG:teuthology.run_tasks:Unwinding manager pcp 2026-03-07T10:13:57.196 DEBUG:teuthology.run_tasks:Unwinding manager internal.timer 2026-03-07T10:13:57.199 INFO:teuthology.task.internal:Duration was 555.759073 seconds 2026-03-07T10:13:57.199 DEBUG:teuthology.run_tasks:Unwinding manager internal.syslog 2026-03-07T10:13:57.201 INFO:teuthology.task.internal.syslog:Shutting down syslog monitoring... 2026-03-07T10:13:57.201 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-07T10:13:57.201 DEBUG:teuthology.orchestra.run.vm10:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-07T10:13:57.221 INFO:teuthology.task.internal.syslog:Checking logs for errors... 2026-03-07T10:13:57.221 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm02.local 2026-03-07T10:13:57.221 DEBUG:teuthology.orchestra.run.vm02:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-07T10:13:57.273 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm10.local 2026-03-07T10:13:57.273 DEBUG:teuthology.orchestra.run.vm10:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-07T10:13:57.282 INFO:teuthology.task.internal.syslog:Gathering journactl... 2026-03-07T10:13:57.282 DEBUG:teuthology.orchestra.run.vm02:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-07T10:13:57.316 DEBUG:teuthology.orchestra.run.vm10:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-07T10:13:57.355 INFO:teuthology.task.internal.syslog:Compressing syslogs... 2026-03-07T10:13:57.355 DEBUG:teuthology.orchestra.run.vm02:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-07T10:13:57.392 DEBUG:teuthology.orchestra.run.vm10:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-07T10:13:57.398 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-07T10:13:57.398 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-07T10:13:57.398 INFO:teuthology.orchestra.run.vm02.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-07T10:13:57.398 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-07T10:13:57.399 INFO:teuthology.orchestra.run.vm02.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: /home/ubuntu/cephtest/archive/syslog/journalctl.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-07T10:13:57.404 INFO:teuthology.orchestra.run.vm10.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-07T10:13:57.404 INFO:teuthology.orchestra.run.vm10.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-07T10:13:57.404 INFO:teuthology.orchestra.run.vm10.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-07T10:13:57.404 INFO:teuthology.orchestra.run.vm10.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-07T10:13:57.405 INFO:teuthology.orchestra.run.vm10.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0%/home/ubuntu/cephtest/archive/syslog/journalctl.log: -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-07T10:13:57.407 INFO:teuthology.orchestra.run.vm02.stderr: 88.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-07T10:13:57.411 INFO:teuthology.orchestra.run.vm10.stderr: 87.9% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-07T10:13:57.412 DEBUG:teuthology.run_tasks:Unwinding manager internal.sudo 2026-03-07T10:13:57.414 INFO:teuthology.task.internal:Restoring /etc/sudoers... 2026-03-07T10:13:57.414 DEBUG:teuthology.orchestra.run.vm02:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-07T10:13:57.459 DEBUG:teuthology.orchestra.run.vm10:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-07T10:13:57.466 DEBUG:teuthology.run_tasks:Unwinding manager internal.coredump 2026-03-07T10:13:57.469 DEBUG:teuthology.orchestra.run.vm02:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-07T10:13:57.504 DEBUG:teuthology.orchestra.run.vm10:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-07T10:13:57.509 INFO:teuthology.orchestra.run.vm02.stdout:kernel.core_pattern = core 2026-03-07T10:13:57.515 INFO:teuthology.orchestra.run.vm10.stdout:kernel.core_pattern = core 2026-03-07T10:13:57.522 DEBUG:teuthology.orchestra.run.vm02:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-07T10:13:57.561 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-07T10:13:57.562 DEBUG:teuthology.orchestra.run.vm10:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-07T10:13:57.566 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-07T10:13:57.566 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive 2026-03-07T10:13:57.569 INFO:teuthology.task.internal:Transferring archived files... 2026-03-07T10:13:57.569 DEBUG:teuthology.misc:Transferring archived files from vm02:/home/ubuntu/cephtest/archive to /archive/irq0-2026-03-07_10:02:54-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps/9/remote/vm02 2026-03-07T10:13:57.569 DEBUG:teuthology.orchestra.run.vm02:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-07T10:13:57.611 DEBUG:teuthology.misc:Transferring archived files from vm10:/home/ubuntu/cephtest/archive to /archive/irq0-2026-03-07_10:02:54-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps/9/remote/vm10 2026-03-07T10:13:57.611 DEBUG:teuthology.orchestra.run.vm10:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-07T10:13:57.618 INFO:teuthology.task.internal:Removing archive directory... 2026-03-07T10:13:57.618 DEBUG:teuthology.orchestra.run.vm02:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-07T10:13:57.656 DEBUG:teuthology.orchestra.run.vm10:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-07T10:13:57.663 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive_upload 2026-03-07T10:13:57.665 INFO:teuthology.task.internal:Not uploading archives. 2026-03-07T10:13:57.665 DEBUG:teuthology.run_tasks:Unwinding manager internal.base 2026-03-07T10:13:57.667 INFO:teuthology.task.internal:Tidying up after the test... 2026-03-07T10:13:57.667 DEBUG:teuthology.orchestra.run.vm02:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-07T10:13:57.700 DEBUG:teuthology.orchestra.run.vm10:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-07T10:13:57.702 INFO:teuthology.orchestra.run.vm02.stdout: 258079 4 drwxr-xr-x 2 ubuntu ubuntu 4096 Mar 7 10:13 /home/ubuntu/cephtest 2026-03-07T10:13:57.707 INFO:teuthology.orchestra.run.vm10.stdout: 258077 4 drwxr-xr-x 2 ubuntu ubuntu 4096 Mar 7 10:13 /home/ubuntu/cephtest 2026-03-07T10:13:57.708 DEBUG:teuthology.run_tasks:Unwinding manager console_log 2026-03-07T10:13:57.713 INFO:teuthology.run:Summary data: description: orch:cephadm:workunits/{0-distro/ubuntu_22.04 agent/off mon_election/classic task/test_extra_daemon_features} duration: 555.7590732574463 owner: irq0 success: true 2026-03-07T10:13:57.713 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-07T10:13:57.731 INFO:teuthology.run:pass